Skip to content

Instantly share code, notes, and snippets.

@ezyang
Created December 12, 2022 18:12
Show Gist options
  • Star 0 You must be signed in to star a gist
  • Fork 0 You must be signed in to fork a gist
  • Save ezyang/6cad3b202daf5c35f93acbb4d71afa5f to your computer and use it in GitHub Desktop.
Save ezyang/6cad3b202daf5c35f93acbb4d71afa5f to your computer and use it in GitHub Desktop.
Sweep logs for HEAD --accuracy --backend inductor --training --explain (TORCHDYNAMO_DYNAMIC_SHAPES=1) - bcb284d77fe865373b2f1617867320fb32ea68af Mon Dec 12 06:20:15 PST 2022
This file has been truncated, but you can view the full file.
cuda train BERT_pytorch ERROR:common:compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 952, in aot_dispatch_base
compiled_fw = aot_config.fw_compiler(fw_module, flat_args)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 437, in codegen
self.init_wrapper_code()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 431, in init_wrapper_code
self.wrapper_code = WrapperCodeGen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 302, in __init__
self.write_prefix()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 335, in write_prefix
V.graph.sizevars.codegen(self.wrapper_call, V.graph.graph_inputs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 494, in codegen
assert not needed
AssertionError
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 332, in wrapper
self.output.compile_subgraph(self, reason=reason)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 5 ops
cuda train Background_Matting [2022-12-12 06:21:23,943] torch._inductor.graph: [WARNING] Creating implicit fallback for:
target: <function sym_float at 0x7f6cde1a2ca0>
args[0]: 256.0
ERROR:common:compile_fx raised LoweringException: TypeError: sym_float() missing 1 required positional argument: 'a'
target: <function sym_float at 0x7f6cde1a2ca0>
args[0]: 256.0
While executing %sym_float : [#users=1] = call_function[target=torch.fx.experimental.symbolic_shapes.sym_float](args = (%mul_108,), kwargs = {})
Original traceback:
None
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 272, in call_function
out = lowerings[target](*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 221, in wrapped
return decomp_fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 991, in handler
TensorBox.create, ir.FallbackKernel.create(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 3011, in create
) = cls.process_kernel(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 2406, in process_kernel
example_output = kernel(*new_args, **new_kwargs)
TypeError: sym_float() missing 1 required positional argument: 'a'
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 275, in call_function
raise LoweringException(e, target, args, kwargs) from e
torch._inductor.exc.LoweringException: TypeError: sym_float() missing 1 required positional argument: 'a'
target: <function sym_float at 0x7f6cde1a2ca0>
args[0]: 256.0
While executing %sym_float : [#users=1] = call_function[target=torch.fx.experimental.symbolic_shapes.sym_float](args = (%mul_108,), kwargs = {})
Original traceback:
None
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised LoweringException: TypeError: sym_float() missing 1 required positional argument: 'a'
target: <function sym_float at 0x7f6cde1a2ca0>
args[0]: 256.0
While executing %sym_float : [#users=1] = call_function[target=torch.fx.experimental.symbolic_shapes.sym_float](args = (%mul_108,), kwargs = {})
Original traceback:
None
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 183 ops
WARNING:root:DALLE2_pytorch failed to load
Eager model failed to run
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 998, in validate_model
self.model_iter_fn(model, example_inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 361, in forward_and_backward_pass
self.grad_scaler.scale(loss).backward()
File "/data/users/ezyang/a/pytorch/torch/_tensor.py", line 484, in backward
torch.autograd.backward(
File "/data/users/ezyang/a/pytorch/torch/autograd/__init__.py", line 197, in backward
Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
RuntimeError: element 0 of tensors does not require grad and does not have a grad_fn
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 2010, in run
) = runner.load_model(device, model_name, batch_size=batch_size)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 305, in load_model
self.validate_model(model, example_inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1000, in validate_model
raise NotImplementedError("Eager model failed to run") from e
NotImplementedError: Eager model failed to run
cuda train LearningToPaint ERROR:common:name 's0' is not defined
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/models/LearningToPaint/baseline/DRL/actor.py", line 104, in forward
def forward(self, x):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2340, in forward
return compiled_fn(full_args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 887, in g
return f(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1903, in debug_compiled_function
return compiled_function(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1718, in compiled_function
all_outs = CompiledFunction.apply(*args_with_synthetic_bases)
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 419, in apply
return super().apply(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1581, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 912, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 199, in run
return model(new_inputs)
File "/tmp/torchinductor_ezyang/mh/cmhdhkraqea5462gm4idaxjnb3czmtuq2j2l2et3rzcnx6p33x7a.py", line 944, in call
return (buf1, buf2, buf5, buf6, buf9, buf10, buf12, buf13, buf17, buf18, buf21, buf22, buf25, buf26, buf29, buf30, buf32, buf33, buf37, buf38, buf41, buf42, buf45, buf46, buf49, buf50, buf52, buf53, buf57, buf58, buf61, buf62, buf65, buf66, buf69, buf70, buf72, buf73, buf77, buf78, buf81, buf82, buf86, primals_1, primals_2, primals_4, primals_5, primals_7, primals_8, primals_10, primals_11, primals_13, primals_14, primals_16, primals_17, primals_19, primals_20, primals_22, primals_23, primals_25, primals_26, primals_28, primals_29, primals_31, primals_32, primals_34, primals_35, primals_37, primals_38, primals_40, primals_41, primals_43, primals_44, primals_46, primals_47, primals_49, primals_50, primals_52, primals_53, primals_55, primals_56, primals_58, primals_59, primals_61, primals_62, primals_129, buf0, buf1, buf2, buf3, buf4, buf5, buf6, buf7, buf8, buf9, buf10, buf11, buf12, buf13, buf15, buf16, buf17, buf18, buf19, buf20, buf21, buf22, buf23, buf24, buf25, buf26, buf27, buf28, buf29, buf30, buf31, buf32, buf33, buf35, buf36, buf37, buf38, buf39, buf40, buf41, buf42, buf43, buf44, buf45, buf46, buf47, buf48, buf49, buf50, buf51, buf52, buf53, buf55, buf56, buf57, buf58, buf59, buf60, buf61, buf62, buf63, buf64, buf65, buf66, buf67, buf68, buf69, buf70, buf71, buf72, buf73, buf75, buf76, buf77, buf78, buf79, buf80, buf81, buf82, buf83, buf84, buf86, as_strided(primals_64, (65, 512), (512, 1)), s0, )
NameError: name 's0' is not defined
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 72 ops
cuda train Super_SloMo ERROR:common:compile_fx raised RuntimeError: Trying to extract a concrete int out of a symbolic int
While executing %primals_113 : [#users=1] = placeholder[target=primals_113]
Original traceback:
None
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/cache.py", line 70, in wrapper
retval = cfunc(*args, **kwargs)
TypeError: unhashable type: 'SymInt'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 231, in placeholder
sizes, strides = self.static_sizes_strides(example)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 62, in static_sizes_strides
size = [sympy.Integer(i) for i in ex.size()]
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 62, in <listcomp>
size = [sympy.Integer(i) for i in ex.size()]
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/cache.py", line 74, in wrapper
retval = func(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/numbers.py", line 2095, in __new__
ival = int(i)
File "/data/users/ezyang/a/pytorch/torch/__init__.py", line 242, in __int__
return self.node.int_()
File "/data/users/ezyang/a/pytorch/torch/fx/experimental/symbolic_shapes.py", line 210, in int_
raise RuntimeError("Trying to extract a concrete int out of a symbolic int")
RuntimeError: Trying to extract a concrete int out of a symbolic int
While executing %primals_113 : [#users=1] = placeholder[target=primals_113]
Original traceback:
None
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised RuntimeError: Trying to extract a concrete int out of a symbolic int
While executing %primals_113 : [#users=1] = placeholder[target=primals_113]
Original traceback:
None
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 374 ops
cuda train alexnet ERROR:common:name 's0' is not defined
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/torchvision/torchvision/models/alexnet.py", line 47, in forward
def forward(self, x: torch.Tensor) -> torch.Tensor:
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2340, in forward
return compiled_fn(full_args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 887, in g
return f(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1903, in debug_compiled_function
return compiled_function(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1718, in compiled_function
all_outs = CompiledFunction.apply(*args_with_synthetic_bases)
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 419, in apply
return super().apply(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1581, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 912, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 199, in run
return model(new_inputs)
File "/tmp/torchinductor_ezyang/m2/cm2m6q3kj6mp7dnzzrlg3nuw7z7vmhps3uhynbgm4zrdsc2dzwjp.py", line 386, in call
return (buf21, primals_1, primals_3, primals_5, primals_7, primals_9, primals_17, buf1, buf2, buf3, buf5, buf6, buf7, buf9, buf11, buf13, buf14, buf15, as_strided(buf16, (4, 9216), (9216, 1)), buf18, buf20, as_strided(primals_15, (1000, 4096), (4096, 1)), as_strided(primals_13, (4096, 4096), (4096, 1)), as_strided(primals_11, (4096, 9216), (9216, 1)), s0, )
NameError: name 's0' is not defined
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 22 ops
cuda train attention_is_all_you_need_pytorch ERROR:common:compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 952, in aot_dispatch_base
compiled_fw = aot_config.fw_compiler(fw_module, flat_args)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 437, in codegen
self.init_wrapper_code()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 431, in init_wrapper_code
self.wrapper_code = WrapperCodeGen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 302, in __init__
self.write_prefix()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 335, in write_prefix
V.graph.sizevars.codegen(self.wrapper_call, V.graph.graph_inputs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 494, in codegen
assert not needed
AssertionError
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 332, in wrapper
self.output.compile_subgraph(self, reason=reason)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 11 ops
cuda train dcgan ERROR:common:'ShapeEnv' object has no attribute 'create_symbolic_sizes_strides'
While executing %tangents_1 : [#users=0] = placeholder[target=tangents_1]
Original traceback:
None
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 361, in <graph break in forward_and_backward_pass>
self.grad_scaler.scale(loss).backward()
File "/data/users/ezyang/a/pytorch/torch/_tensor.py", line 484, in backward
torch.autograd.backward(
File "/data/users/ezyang/a/pytorch/torch/autograd/__init__.py", line 197, in backward
Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 272, in apply
return user_fn(self, *args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1683, in backward
CompiledFunction.compiled_bw = aot_config.bw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 68, in _wrapped_bw_compiler
return eval_frame.disable(eval_frame.disable(bw_compiler)(*args, **kwargs))
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 380, in bw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 233, in placeholder
sizes, strides = self.symbolic_sizes_strides(example)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 50, in symbolic_sizes_strides
size, stride = self._shape_env.create_symbolic_sizes_strides(ex)
AttributeError: 'ShapeEnv' object has no attribute 'create_symbolic_sizes_strides'
While executing %tangents_1 : [#users=0] = placeholder[target=tangents_1]
Original traceback:
None
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 13 ops
cuda train densenet121 ERROR:common:name 's0' is not defined
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/torchvision/torchvision/models/densenet.py", line 212, in forward
def forward(self, x: Tensor) -> Tensor:
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2340, in forward
return compiled_fn(full_args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 887, in g
return f(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1903, in debug_compiled_function
return compiled_function(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1718, in compiled_function
all_outs = CompiledFunction.apply(*args_with_synthetic_bases)
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 419, in apply
return super().apply(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1581, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 912, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 199, in run
return model(new_inputs)
File "/tmp/torchinductor_ezyang/ed/cedphlhn6ebmmibptrxt4otj3zfyv4zj57m3eonhhak6fd6ulh45.py", line 6557, in call
return (buf1, buf2, buf6, buf7, buf10, buf11, buf16, buf17, buf20, buf21, buf28, buf29, buf32, buf33, buf41, buf42, buf45, buf46, buf55, buf56, buf59, buf60, buf70, buf71, buf74, buf75, buf86, buf87, buf91, buf92, buf95, buf96, buf101, buf102, buf105, buf106, buf113, buf114, buf117, buf118, buf126, buf127, buf130, buf131, buf140, buf141, buf144, buf145, buf155, buf156, buf159, buf160, buf171, buf172, buf175, buf176, buf188, buf189, buf192, buf193, buf206, buf207, buf210, buf211, buf225, buf226, buf229, buf230, buf245, buf246, buf249, buf250, buf266, buf267, buf270, buf271, buf288, buf289, buf293, buf294, buf297, buf298, buf303, buf304, buf307, buf308, buf315, buf316, buf319, buf320, buf328, buf329, buf332, buf333, buf342, buf343, buf346, buf347, buf357, buf358, buf361, buf362, buf373, buf374, buf377, buf378, buf390, buf391, buf394, buf395, buf408, buf409, buf412, buf413, buf427, buf428, buf431, buf432, buf447, buf448, buf451, buf452, buf468, buf469, buf472, buf473, buf490, buf491, buf494, buf495, buf513, buf514, buf517, buf518, buf537, buf538, buf541, buf542, buf562, buf563, buf566, buf567, buf588, buf589, buf592, buf593, buf615, buf616, buf619, buf620, buf643, buf644, buf647, buf648, buf672, buf673, buf676, buf677, buf702, buf703, buf706, buf707, buf733, buf734, buf737, buf738, buf765, buf766, buf769, buf770, buf798, buf799, buf802, buf803, buf832, buf833, buf837, buf838, buf841, buf842, buf847, buf848, buf851, buf852, buf859, buf860, buf863, buf864, buf872, buf873, buf876, buf877, buf886, buf887, buf890, buf891, buf901, buf902, buf905, buf906, buf917, buf918, buf921, buf922, buf934, buf935, buf938, buf939, buf952, buf953, buf956, buf957, buf971, buf972, buf975, buf976, buf991, buf992, buf995, buf996, buf1012, buf1013, buf1016, buf1017, buf1034, buf1035, buf1038, buf1039, buf1057, buf1058, buf1061, buf1062, buf1081, buf1082, buf1085, buf1086, buf1106, buf1107, buf1110, buf1111, buf1132, buf1133, buf1137, primals_1, primals_2, primals_4, primals_6, primals_7, primals_9, primals_10, primals_12, primals_13, primals_15, primals_16, primals_18, primals_19, primals_21, primals_22, primals_24, primals_25, primals_27, primals_28, primals_30, primals_31, primals_33, primals_34, primals_36, primals_37, primals_39, primals_40, primals_42, primals_43, primals_45, primals_46, primals_48, primals_49, primals_51, primals_52, primals_54, primals_55, primals_57, primals_58, primals_60, primals_61, primals_63, primals_64, primals_66, primals_67, primals_69, primals_70, primals_72, primals_73, primals_75, primals_76, primals_78, primals_79, primals_81, primals_82, primals_84, primals_85, primals_87, primals_88, primals_90, primals_91, primals_93, primals_94, primals_96, primals_97, primals_99, primals_100, primals_102, primals_103, primals_105, primals_106, primals_108, primals_109, primals_111, primals_112, primals_114, primals_115, primals_117, primals_118, primals_120, primals_121, primals_123, primals_124, primals_126, primals_127, primals_129, primals_130, primals_132, primals_133, primals_135, primals_136, primals_138, primals_139, primals_141, primals_142, primals_144, primals_145, primals_147, primals_148, primals_150, primals_151, primals_153, primals_154, primals_156, primals_157, primals_159, primals_160, primals_162, primals_163, primals_165, primals_166, primals_168, primals_169, primals_171, primals_172, primals_174, primals_175, primals_177, primals_178, primals_180, primals_181, primals_183, primals_184, primals_186, primals_187, primals_189, primals_190, primals_192, primals_193, primals_195, primals_196, primals_198, primals_199, primals_201, primals_202, primals_204, primals_205, primals_207, primals_208, primals_210, primals_211, primals_213, primals_214, primals_216, primals_217, primals_219, primals_220, primals_222, primals_223, primals_225, primals_226, primals_228, primals_229, primals_231, primals_232, primals_234, primals_235, primals_237, primals_238, primals_240, primals_241, primals_243, primals_244, primals_246, primals_247, primals_249, primals_250, primals_252, primals_253, primals_255, primals_256, primals_258, primals_259, primals_261, primals_262, primals_264, primals_265, primals_267, primals_268, primals_270, primals_271, primals_273, primals_274, primals_276, primals_277, primals_279, primals_280, primals_282, primals_283, primals_285, primals_286, primals_288, primals_289, primals_291, primals_292, primals_294, primals_295, primals_297, primals_298, primals_300, primals_301, primals_303, primals_304, primals_306, primals_307, primals_309, primals_310, primals_312, primals_313, primals_315, primals_316, primals_318, primals_319, primals_321, primals_322, primals_324, primals_325, primals_327, primals_328, primals_330, primals_331, primals_333, primals_334, primals_336, primals_337, primals_339, primals_340, primals_342, primals_343, primals_345, primals_346, primals_348, primals_349, primals_351, primals_352, primals_354, primals_355, primals_357, primals_358, primals_360, primals_361, primals_728, buf0, buf1, buf2, buf3, buf5, buf4, buf6, buf7, buf8, buf9, buf10, buf11, buf12, buf15, buf16, buf17, buf18, buf19, buf20, buf21, buf22, buf27, buf28, buf29, buf30, buf31, buf32, buf33, buf34, buf40, buf41, buf42, buf43, buf44, buf45, buf46, buf47, buf54, buf55, buf56, buf57, buf58, buf59, buf60, buf61, buf69, buf70, buf71, buf72, buf73, buf74, buf75, buf76, buf85, buf86, buf87, buf88, buf89, buf90, buf91, buf92, buf93, buf94, buf95, buf96, buf97, buf100, buf101, buf102, buf103, buf104, buf105, buf106, buf107, buf112, buf113, buf114, buf115, buf116, buf117, buf118, buf119, buf125, buf126, buf127, buf128, buf129, buf130, buf131, buf132, buf139, buf140, buf141, buf142, buf143, buf144, buf145, buf146, buf154, buf155, buf156, buf157, buf158, buf159, buf160, buf161, buf170, buf171, buf172, buf173, buf174, buf175, buf176, buf177, buf187, buf188, buf189, buf190, buf191, buf192, buf193, buf194, buf205, buf206, buf207, buf208, buf209, buf210, buf211, buf212, buf224, buf225, buf226, buf227, buf228, buf229, buf230, buf231, buf244, buf245, buf246, buf247, buf248, buf249, buf250, buf251, buf265, buf266, buf267, buf268, buf269, buf270, buf271, buf272, buf287, buf288, buf289, buf290, buf291, buf292, buf293, buf294, buf295, buf296, buf297, buf298, buf299, buf302, buf303, buf304, buf305, buf306, buf307, buf308, buf309, buf314, buf315, buf316, buf317, buf318, buf319, buf320, buf321, buf327, buf328, buf329, buf330, buf331, buf332, buf333, buf334, buf341, buf342, buf343, buf344, buf345, buf346, buf347, buf348, buf356, buf357, buf358, buf359, buf360, buf361, buf362, buf363, buf372, buf373, buf374, buf375, buf376, buf377, buf378, buf379, buf389, buf390, buf391, buf392, buf393, buf394, buf395, buf396, buf407, buf408, buf409, buf410, buf411, buf412, buf413, buf414, buf426, buf427, buf428, buf429, buf430, buf431, buf432, buf433, buf446, buf447, buf448, buf449, buf450, buf451, buf452, buf453, buf467, buf468, buf469, buf470, buf471, buf472, buf473, buf474, buf489, buf490, buf491, buf492, buf493, buf494, buf495, buf496, buf512, buf513, buf514, buf515, buf516, buf517, buf518, buf519, buf536, buf537, buf538, buf539, buf540, buf541, buf542, buf543, buf561, buf562, buf563, buf564, buf565, buf566, buf567, buf568, buf587, buf588, buf589, buf590, buf591, buf592, buf593, buf594, buf614, buf615, buf616, buf617, buf618, buf619, buf620, buf621, buf642, buf643, buf644, buf645, buf646, buf647, buf648, buf649, buf671, buf672, buf673, buf674, buf675, buf676, buf677, buf678, buf701, buf702, buf703, buf704, buf705, buf706, buf707, buf708, buf732, buf733, buf734, buf735, buf736, buf737, buf738, buf739, buf764, buf765, buf766, buf767, buf768, buf769, buf770, buf771, buf797, buf798, buf799, buf800, buf801, buf802, buf803, buf804, buf831, buf832, buf833, buf834, buf835, buf836, buf837, buf838, buf839, buf840, buf841, buf842, buf843, buf846, buf847, buf848, buf849, buf850, buf851, buf852, buf853, buf858, buf859, buf860, buf861, buf862, buf863, buf864, buf865, buf871, buf872, buf873, buf874, buf875, buf876, buf877, buf878, buf885, buf886, buf887, buf888, buf889, buf890, buf891, buf892, buf900, buf901, buf902, buf903, buf904, buf905, buf906, buf907, buf916, buf917, buf918, buf919, buf920, buf921, buf922, buf923, buf933, buf934, buf935, buf936, buf937, buf938, buf939, buf940, buf951, buf952, buf953, buf954, buf955, buf956, buf957, buf958, buf970, buf971, buf972, buf973, buf974, buf975, buf976, buf977, buf990, buf991, buf992, buf993, buf994, buf995, buf996, buf997, buf1011, buf1012, buf1013, buf1014, buf1015, buf1016, buf1017, buf1018, buf1033, buf1034, buf1035, buf1036, buf1037, buf1038, buf1039, buf1040, buf1056, buf1057, buf1058, buf1059, buf1060, buf1061, buf1062, buf1063, buf1080, buf1081, buf1082, buf1083, buf1084, buf1085, buf1086, buf1087, buf1105, buf1106, buf1107, buf1108, buf1109, buf1110, buf1111, buf1112, buf1131, buf1132, buf1133, as_strided(buf1136, (4, 1024), (1024, 1)), as_strided(primals_363, (1000, 1024), (1024, 1)), buf1138, s0, 7, 7, )
NameError: name 's0' is not defined
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 431 ops
WARNING:root:detectron2_fcos_r_50_fpn failed to load
FCOS train is not supported by upstream detectron2. See GH Issue: https://github.com/facebookresearch/detectron2/issues/4369.
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 2010, in run
) = runner.load_model(device, model_name, batch_size=batch_size)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 268, in load_model
benchmark = benchmark_cls(
File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/util/model.py", line 18, in __call__
obj = type.__call__(cls, *args, **kwargs)
File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/models/detectron2_fcos_r_50_fpn/__init__.py", line 15, in __init__
super().__init__(variant="COCO-Detection/fcos_R_50_FPN_1x.py", test=test, device=device,
File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/util/framework/detectron2/model_factory.py", line 100, in __init__
loader = self.setup_train(cfg, args)
File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/util/framework/detectron2/model_factory.py", line 110, in setup_train
raise NotImplementedError("FCOS train is not supported by upstream detectron2. " \
NotImplementedError: FCOS train is not supported by upstream detectron2. See GH Issue: https://github.com/facebookresearch/detectron2/issues/4369.
WARNING:root:detectron2_maskrcnn_r_50_c4 failed to load
Eager model failed to run
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 998, in validate_model
self.model_iter_fn(model, example_inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 360, in forward_and_backward_pass
loss = self.compute_loss(pred)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 350, in compute_loss
return reduce_to_scalar_loss(pred)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/testing.py", line 97, in reduce_to_scalar_loss
return sum([reduce_to_scalar_loss(x) for x in out]) / len(out)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/testing.py", line 97, in <listcomp>
return sum([reduce_to_scalar_loss(x) for x in out]) / len(out)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/testing.py", line 107, in reduce_to_scalar_loss
return sum([reduce_to_scalar_loss(value) for value in out.values()]) / len(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/testing.py", line 107, in <listcomp>
return sum([reduce_to_scalar_loss(value) for value in out.values()]) / len(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/testing.py", line 110, in reduce_to_scalar_loss
raise NotImplementedError("Don't know how to reduce", type(out))
NotImplementedError: ("Don't know how to reduce", <class 'detectron2.structures.instances.Instances'>)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 2010, in run
) = runner.load_model(device, model_name, batch_size=batch_size)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 305, in load_model
self.validate_model(model, example_inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1000, in validate_model
raise NotImplementedError("Eager model failed to run") from e
NotImplementedError: Eager model failed to run
cuda train dlrm ERROR:common:compile_fx raised NotImplementedError: Cannot access storage of SparseTensorImpl
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1494, in aot_dispatch_autograd
fx_g = make_fx(joint_forward_backward, aot_config.decompositions)(
File "/data/users/ezyang/a/pytorch/torch/fx/experimental/proxy_tensor.py", line 691, in wrapped
t = dispatch_trace(wrap_key(func, args, fx_tracer), tracer=fx_tracer, concrete_args=tuple(phs))
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/fx/experimental/proxy_tensor.py", line 441, in dispatch_trace
graph = tracer.trace(root, concrete_args)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/fx/_symbolic_trace.py", line 756, in trace
(self.create_arg(fn(*args)),),
File "/data/users/ezyang/a/pytorch/torch/fx/_symbolic_trace.py", line 630, in flatten_fn
tree_out = root_fn(*tree_args)
File "/data/users/ezyang/a/pytorch/torch/fx/experimental/proxy_tensor.py", line 457, in wrapped
out = f(*tensors)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 822, in functionalized_joint
outs = joint_forward_backward(f_primals, f_tangents)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 789, in joint_forward_backward
backward_out = torch.autograd.grad(
File "/data/users/ezyang/a/pytorch/torch/autograd/__init__.py", line 266, in grad
return handle_torch_function(
File "/data/users/ezyang/a/pytorch/torch/overrides.py", line 1520, in handle_torch_function
result = mode.__torch_function__(public_api, types, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/overrides.py", line 37, in __torch_function__
return func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/autograd/__init__.py", line 300, in grad
return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
NotImplementedError: Cannot access storage of SparseTensorImpl
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 476, in compile_subgraph
self.compile_and_call_fx_graph(tx, list(reversed(stack_values)), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised NotImplementedError: Cannot access storage of SparseTensorImpl
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 40 ops
/data/users/ezyang/a/pytorch/torch/utils/tensorboard/__init__.py:4: DeprecationWarning: distutils Version classes are deprecated. Use packaging.version instead.
if not hasattr(tensorboard, "__version__") or LooseVersion(
/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/gym/core.py:317: DeprecationWarning: WARN: Initializing wrapper in old step API which returns one bool instead of two. It is recommended to set `new_step_api=True` to use new step API. This will be the default behaviour in future.
deprecation(
cuda train drq ERROR:common:'ShapeEnv' object has no attribute 'create_symbolic_sizes_strides'
While executing %tangents_1 : [#users=1] = placeholder[target=tangents_1]
Original traceback:
None
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 361, in <graph break in forward_and_backward_pass>
self.grad_scaler.scale(loss).backward()
File "/data/users/ezyang/a/pytorch/torch/_tensor.py", line 484, in backward
torch.autograd.backward(
File "/data/users/ezyang/a/pytorch/torch/autograd/__init__.py", line 197, in backward
Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 272, in apply
return user_fn(self, *args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1683, in backward
CompiledFunction.compiled_bw = aot_config.bw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 68, in _wrapped_bw_compiler
return eval_frame.disable(eval_frame.disable(bw_compiler)(*args, **kwargs))
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 380, in bw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 233, in placeholder
sizes, strides = self.symbolic_sizes_strides(example)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 50, in symbolic_sizes_strides
size, stride = self._shape_env.create_symbolic_sizes_strides(ex)
AttributeError: 'ShapeEnv' object has no attribute 'create_symbolic_sizes_strides'
While executing %tangents_1 : [#users=1] = placeholder[target=tangents_1]
Original traceback:
None
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 3 graph(s) covering 34 ops
cuda train fastNLP_Bert ERROR:common:compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 952, in aot_dispatch_base
compiled_fw = aot_config.fw_compiler(fw_module, flat_args)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 437, in codegen
self.init_wrapper_code()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 431, in init_wrapper_code
self.wrapper_code = WrapperCodeGen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 302, in __init__
self.write_prefix()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 335, in write_prefix
V.graph.sizevars.codegen(self.wrapper_call, V.graph.graph_inputs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 494, in codegen
assert not needed
AssertionError
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/fastNLP/models/bert.py", line 265, in forward
sequence_output = self.bert(words)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/fastNLP/embeddings/bert_embedding.py", line 137, in forward
outputs = self.model(words)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 332, in wrapper
self.output.compile_subgraph(self, reason=reason)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 13 ops
cuda train functorch_dp_cifar10 ERROR:common:name 's0' is not defined
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/torchvision/torchvision/models/resnet.py", line 284, in forward
def forward(self, x: Tensor) -> Tensor:
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2340, in forward
return compiled_fn(full_args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 887, in g
return f(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1903, in debug_compiled_function
return compiled_function(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1718, in compiled_function
all_outs = CompiledFunction.apply(*args_with_synthetic_bases)
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 419, in apply
return super().apply(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1581, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 912, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 199, in run
return model(new_inputs)
File "/tmp/torchinductor_ezyang/lj/cljg56pbqgfgxf7ybhk5sr2tdonkcbp33cyblginj3jibxqstila.py", line 1185, in call
return (buf105, primals_1, primals_2, primals_4, primals_5, primals_7, primals_8, primals_10, primals_11, primals_13, primals_14, primals_16, primals_17, primals_19, primals_20, primals_22, primals_23, primals_25, primals_26, primals_28, primals_29, primals_31, primals_32, primals_34, primals_35, primals_37, primals_38, primals_40, primals_41, primals_43, primals_44, primals_46, primals_47, primals_49, primals_50, primals_52, primals_53, primals_55, primals_56, primals_58, primals_59, primals_63, buf0, as_strided(buf106, (4, 32), (32, 1)), as_strided(buf107, (4, 32), (32, 1)), buf4, buf5, buf6, buf7, as_strided(buf108, (4, 32), (32, 1)), as_strided(buf109, (4, 32), (32, 1)), buf11, buf12, as_strided(buf110, (4, 32), (32, 1)), as_strided(buf111, (4, 32), (32, 1)), buf16, buf17, as_strided(buf112, (4, 32), (32, 1)), as_strided(buf113, (4, 32), (32, 1)), buf21, buf22, as_strided(buf114, (4, 32), (32, 1)), as_strided(buf115, (4, 32), (32, 1)), buf26, buf27, as_strided(buf116, (4, 32), (32, 1)), as_strided(buf117, (4, 32), (32, 1)), buf31, buf32, as_strided(buf118, (4, 32), (32, 1)), as_strided(buf119, (4, 32), (32, 1)), buf36, as_strided(buf120, (4, 32), (32, 1)), as_strided(buf121, (4, 32), (32, 1)), buf41, buf42, as_strided(buf122, (4, 32), (32, 1)), as_strided(buf123, (4, 32), (32, 1)), buf46, buf47, as_strided(buf124, (4, 32), (32, 1)), as_strided(buf125, (4, 32), (32, 1)), buf51, buf52, as_strided(buf126, (4, 32), (32, 1)), as_strided(buf127, (4, 32), (32, 1)), buf56, buf57, as_strided(buf128, (4, 32), (32, 1)), as_strided(buf129, (4, 32), (32, 1)), buf61, as_strided(buf130, (4, 32), (32, 1)), as_strided(buf131, (4, 32), (32, 1)), buf66, buf67, as_strided(buf132, (4, 32), (32, 1)), as_strided(buf133, (4, 32), (32, 1)), buf71, buf72, as_strided(buf134, (4, 32), (32, 1)), as_strided(buf135, (4, 32), (32, 1)), buf76, buf77, as_strided(buf136, (4, 32), (32, 1)), as_strided(buf137, (4, 32), (32, 1)), buf81, buf82, as_strided(buf138, (4, 32), (32, 1)), as_strided(buf139, (4, 32), (32, 1)), buf86, as_strided(buf140, (4, 32), (32, 1)), as_strided(buf141, (4, 32), (32, 1)), buf91, buf92, as_strided(buf142, (4, 32), (32, 1)), as_strided(buf143, (4, 32), (32, 1)), buf96, buf97, buf101, buf102, as_strided(buf104, (4, 512), (512, 1)), as_strided(primals_61, (1000, 512), (512, 1)), buf144, s0, 16, 16, 256, 8, 8, 64, 8, 8, 64, 8, 8, 64, 8, 8, 64, 4, 4, 16, 4, 4, 16, 4, 4, 16, 4, 4, 16, 4, 4, 16, 2, 2, 4, 2, 2, 4, 2, 2, 4, 2, 2, 4, 2, 2, 4, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, )
NameError: name 's0' is not defined
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 69 ops
cuda train functorch_maml_omniglot ERROR:common:name 's0' is not defined
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/container.py", line 202, in forward
def forward(self, input):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2340, in forward
return compiled_fn(full_args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 887, in g
return f(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1903, in debug_compiled_function
return compiled_function(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1718, in compiled_function
all_outs = CompiledFunction.apply(*args_with_synthetic_bases)
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 419, in apply
return super().apply(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1581, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 912, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 199, in run
return model(new_inputs)
File "/tmp/torchinductor_ezyang/65/c65vfsttcdarnn6xnicpy4ochel42nvco2kf73ju2lz63hq6kt6u.py", line 543, in call
return (buf21, primals_1, primals_3, primals_5, primals_7, primals_9, primals_11, primals_15, buf1, as_strided(buf22, (64, ), (1, )), buf5, buf6, buf23, buf8, as_strided(buf24, (64, ), (1, )), buf12, buf13, buf25, buf15, as_strided(buf26, (64, ), (1, )), buf19, buf27, as_strided(buf20, (5, 64), (64, 1)), as_strided(primals_13, (5, 64), (64, 1)), as_strided(buf28, (1, 64, 1, 1), (0, 1, 0, 0)), as_strided(buf29, (1, 64, 1, 1), (0, 1, 0, 0)), as_strided(buf30, (1, 64, 1, 1), (0, 1, 0, 0)), s0, 1, 1, )
NameError: name 's0' is not defined
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 14 ops
cuda train hf_Albert ERROR:common:compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 440, in codegen
self.scheduler.codegen()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/scheduler.py", line 1129, in codegen
self.get_backend(device).codegen_nodes(node.get_nodes())
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1284, in codegen_nodes
return self.codegen_node_schedule(node_schedule, numel, rnumel)
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1336, in codegen_node_schedule
src_code = kernel.codegen_kernel()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1035, in codegen_kernel
"signature": dict(enumerate(map(signature_of, signature))),
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 57, in signature_of
return JITFunction._key_of(V.graph.sizevars.size_hint(arg.expr))
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 367, in size_hint
return int(out)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 990, in forward
outputs = self.albert(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 737, in forward
encoder_outputs = self.encoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 2 graph(s) covering 560 ops
cuda train hf_Bart ERROR:common:compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 440, in codegen
self.scheduler.codegen()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/scheduler.py", line 1129, in codegen
self.get_backend(device).codegen_nodes(node.get_nodes())
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1284, in codegen_nodes
return self.codegen_node_schedule(node_schedule, numel, rnumel)
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1336, in codegen_node_schedule
src_code = kernel.codegen_kernel()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1035, in codegen_kernel
"signature": dict(enumerate(map(signature_of, signature))),
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 57, in signature_of
return JITFunction._key_of(V.graph.sizevars.size_hint(arg.expr))
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 367, in size_hint
return int(out)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/util/framework/huggingface/model_factory.py", line 41, in forward
return self.model(input_ids=input_ids, decoder_input_ids=decoder_input_ids)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1353, in forward
outputs = self.model(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1222, in forward
encoder_outputs = self.encoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 846, in forward
layer_outputs = encoder_layer(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 323, in forward
hidden_states, attn_weights, _ = self.self_attn(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 516, in step
self.output.compile_subgraph(self, partial_convert=True)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 26 ops
cuda train hf_Bert ERROR:common:compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 440, in codegen
self.scheduler.codegen()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/scheduler.py", line 1129, in codegen
self.get_backend(device).codegen_nodes(node.get_nodes())
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1284, in codegen_nodes
return self.codegen_node_schedule(node_schedule, numel, rnumel)
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1336, in codegen_node_schedule
src_code = kernel.codegen_kernel()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1035, in codegen_kernel
"signature": dict(enumerate(map(signature_of, signature))),
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 57, in signature_of
return JITFunction._key_of(V.graph.sizevars.size_hint(arg.expr))
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 367, in size_hint
return int(out)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1351, in forward
outputs = self.bert(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1018, in forward
encoder_outputs = self.encoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 3 graph(s) covering 552 ops
cuda train hf_BigBird ERROR:common:compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 952, in aot_dispatch_base
compiled_fw = aot_config.fw_compiler(fw_module, flat_args)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 437, in codegen
self.init_wrapper_code()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 431, in init_wrapper_code
self.wrapper_code = WrapperCodeGen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 302, in __init__
self.write_prefix()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 335, in write_prefix
V.graph.sizevars.codegen(self.wrapper_call, V.graph.graph_inputs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 494, in codegen
assert not needed
AssertionError
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2462, in forward
outputs = self.bert(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 516, in step
self.output.compile_subgraph(self, partial_convert=True)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 5 ops
cuda train hf_DistilBert ERROR:common:compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 440, in codegen
self.scheduler.codegen()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/scheduler.py", line 1129, in codegen
self.get_backend(device).codegen_nodes(node.get_nodes())
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1284, in codegen_nodes
return self.codegen_node_schedule(node_schedule, numel, rnumel)
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1336, in codegen_node_schedule
src_code = kernel.codegen_kernel()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1035, in codegen_kernel
"signature": dict(enumerate(map(signature_of, signature))),
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 57, in signature_of
return JITFunction._key_of(V.graph.sizevars.size_hint(arg.expr))
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 367, in size_hint
return int(out)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 649, in forward
dlbrt_output = self.distilbert(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 566, in forward
inputs_embeds = self.embeddings(input_ids) # (bs, seq_length, dim)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 2 graph(s) covering 213 ops
cuda train hf_GPT2 ERROR:common:compile_fx raised AssertionError: s1 is needed but not added
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 437, in codegen
self.init_wrapper_code()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 431, in init_wrapper_code
self.wrapper_code = WrapperCodeGen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 302, in __init__
self.write_prefix()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 335, in write_prefix
V.graph.sizevars.codegen(self.wrapper_call, V.graph.graph_inputs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 493, in codegen
assert shape in added, f"{shape} is needed but not added"
AssertionError: s1 is needed but not added
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1048, in forward
transformer_outputs = self.transformer(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 891, in forward
outputs = block(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 391, in forward
attn_outputs = self.attn(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 332, in forward
attn_output, attn_weights = self._attn(query, key, value, attention_mask, head_mask)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 516, in step
self.output.compile_subgraph(self, partial_convert=True)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 476, in compile_subgraph
self.compile_and_call_fx_graph(tx, list(reversed(stack_values)), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised AssertionError: s1 is needed but not added
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 2 graph(s) covering 33 ops
cuda train hf_GPT2_large PASS
Dynamo produced 0 graph(s) covering 0 ops
cuda train hf_Longformer [2022-12-12 06:41:50,594] torch._inductor.ir: [WARNING] Using FallbackKernel: aten.cumsum
ERROR:common:compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 952, in aot_dispatch_base
compiled_fw = aot_config.fw_compiler(fw_module, flat_args)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 437, in codegen
self.init_wrapper_code()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 431, in init_wrapper_code
self.wrapper_code = WrapperCodeGen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 302, in __init__
self.write_prefix()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 335, in write_prefix
V.graph.sizevars.codegen(self.wrapper_call, V.graph.graph_inputs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 494, in codegen
assert not needed
AssertionError
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1813, in forward
outputs = self.longformer(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1696, in forward
padding_len, input_ids, attention_mask, token_type_ids, position_ids, inputs_embeds = self._pad_to_window_size(
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1715, in <graph break in forward>
encoder_outputs = self.encoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 332, in wrapper
self.output.compile_subgraph(self, reason=reason)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 2 graph(s) covering 30 ops
cuda train hf_Reformer [2022-12-12 06:42:02,721] torch._inductor.ir: [WARNING] Using FallbackKernel: aten.native_dropout
[2022-12-12 06:42:02,724] torch._inductor.graph: [WARNING] Creating implicit fallback for:
target: aten.rand_like.default
args[0]: TensorBox(StorageBox(
Pointwise(
'cuda',
torch.float32,
constant(0, torch.float32),
ranges=[s0, 64, 1, 1],
origins={empty}
)
))
kwargs: {'dtype': torch.float32, 'layout': torch.strided, 'device': device(type='cuda', index=0), 'pin_memory': False}
[2022-12-12 06:42:02,729] torch._inductor.ir: [WARNING] Using FallbackKernel: torch.ops.aten.rand_like.default
ERROR:common:name 's0' is not defined
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 2397, in forward
reformer_outputs = self.reformer(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 2063, in forward
least_common_mult_chunk_length = _get_least_common_mult_chunk_len(self.config)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 2100, in <graph break in forward>
embedding_output = self.embeddings(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 239, in forward
def forward(self, input_ids=None, position_ids=None, inputs_embeds=None, start_idx_pos_encodings=0):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2340, in forward
return compiled_fn(full_args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 887, in g
return f(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1903, in debug_compiled_function
return compiled_function(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1718, in compiled_function
all_outs = CompiledFunction.apply(*args_with_synthetic_bases)
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 419, in apply
return super().apply(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1581, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 912, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 199, in run
return model(new_inputs)
File "/tmp/torchinductor_ezyang/cu/ccu6uydvpuo5fmjs2tzpghbvbsqwlt63k7pgemkgfr6f3dt6ipj2.py", line 218, in call
return (buf11, buf3, buf10, buf12, s0, )
NameError: name 's0' is not defined
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 2 graph(s) covering 33 ops
cuda train hf_T5 WARNING:common:fp64 golden ref were not generated for hf_T5. Setting accuracy check to cosine
ERROR:common:compile_fx raised LoweringException: RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s1
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s1 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
target: aten.arange.default
args[0]: s1
kwargs: {'dtype': torch.int64, 'device': device(type='cuda', index=0), 'pin_memory': False}
While executing %arange : [#users=2] = call_function[target=torch.ops.aten.arange.default](args = (%sym_size,), kwargs = {dtype: torch.int64, device: cuda:0, pin_memory: False})
Original traceback:
Module stack: {'self_model': "<class 'transformers.models.t5.modeling_t5.T5ForConditionalGeneration'>", 'self_model_encoder': "<class 'transformers.models.t5.modeling_t5.T5Stack'>", 'self_model_encoder_block_0': "<class 'transformers.models.t5.modeling_t5.T5Block'>", 'sub0_0': "<class 'transformers.models.t5.modeling_t5.T5LayerSelfAttention'>", 'self_model_encoder_block_0_layer_0_SelfAttention': "<class 'transformers.models.t5.modeling_t5.T5Attention'>"}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 423, in compute_bias
context_position = torch.arange(query_length, dtype=torch.long, device=device)[:, None]
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 519, in forward
position_bias = self.compute_bias(real_seq_length, key_length, device=scores.device)
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 570, in forward
attention_output = self.SelfAttention(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 664, in forward
self_attention_outputs = self.layer[0](
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1033, in forward
layer_outputs = layer_module(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1601, in forward
encoder_outputs = self.encoder(
| File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/util/framework/huggingface/model_factory.py", line 41, in forward
return self.model(input_ids=input_ids, decoder_input_ids=decoder_input_ids)
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 272, in call_function
out = lowerings[target](*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 221, in wrapped
return decomp_fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 1318, in arange
return fallback_arange(
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 991, in handler
TensorBox.create, ir.FallbackKernel.create(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 3011, in create
) = cls.process_kernel(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 2406, in process_kernel
example_output = kernel(*new_args, **new_kwargs)
File "/data/users/ezyang/a/pytorch/torch/_ops.py", line 500, in __call__
return self._op(*args, **kwargs or {})
File "/data/users/ezyang/a/pytorch/torch/_inductor/overrides.py", line 37, in __torch_function__
return func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_ops.py", line 500, in __call__
return self._op(*args, **kwargs or {})
RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s1
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s1 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 275, in call_function
raise LoweringException(e, target, args, kwargs) from e
torch._inductor.exc.LoweringException: RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s1
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s1 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
target: aten.arange.default
args[0]: s1
kwargs: {'dtype': torch.int64, 'device': device(type='cuda', index=0), 'pin_memory': False}
While executing %arange : [#users=2] = call_function[target=torch.ops.aten.arange.default](args = (%sym_size,), kwargs = {dtype: torch.int64, device: cuda:0, pin_memory: False})
Original traceback:
Module stack: {'self_model': "<class 'transformers.models.t5.modeling_t5.T5ForConditionalGeneration'>", 'self_model_encoder': "<class 'transformers.models.t5.modeling_t5.T5Stack'>", 'self_model_encoder_block_0': "<class 'transformers.models.t5.modeling_t5.T5Block'>", 'sub0_0': "<class 'transformers.models.t5.modeling_t5.T5LayerSelfAttention'>", 'self_model_encoder_block_0_layer_0_SelfAttention': "<class 'transformers.models.t5.modeling_t5.T5Attention'>"}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 423, in compute_bias
context_position = torch.arange(query_length, dtype=torch.long, device=device)[:, None]
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 519, in forward
position_bias = self.compute_bias(real_seq_length, key_length, device=scores.device)
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 570, in forward
attention_output = self.SelfAttention(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 664, in forward
self_attention_outputs = self.layer[0](
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1033, in forward
layer_outputs = layer_module(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1601, in forward
encoder_outputs = self.encoder(
| File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/util/framework/huggingface/model_factory.py", line 41, in forward
return self.model(input_ids=input_ids, decoder_input_ids=decoder_input_ids)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised LoweringException: RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s1
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s1 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
target: aten.arange.default
args[0]: s1
kwargs: {'dtype': torch.int64, 'device': device(type='cuda', index=0), 'pin_memory': False}
While executing %arange : [#users=2] = call_function[target=torch.ops.aten.arange.default](args = (%sym_size,), kwargs = {dtype: torch.int64, device: cuda:0, pin_memory: False})
Original traceback:
Module stack: {'self_model': "<class 'transformers.models.t5.modeling_t5.T5ForConditionalGeneration'>", 'self_model_encoder': "<class 'transformers.models.t5.modeling_t5.T5Stack'>", 'self_model_encoder_block_0': "<class 'transformers.models.t5.modeling_t5.T5Block'>", 'sub0_0': "<class 'transformers.models.t5.modeling_t5.T5LayerSelfAttention'>", 'self_model_encoder_block_0_layer_0_SelfAttention': "<class 'transformers.models.t5.modeling_t5.T5Attention'>"}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 423, in compute_bias
context_position = torch.arange(query_length, dtype=torch.long, device=device)[:, None]
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 519, in forward
position_bias = self.compute_bias(real_seq_length, key_length, device=scores.device)
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 570, in forward
attention_output = self.SelfAttention(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 664, in forward
self_attention_outputs = self.layer[0](
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1033, in forward
layer_outputs = layer_module(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1601, in forward
encoder_outputs = self.encoder(
| File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/util/framework/huggingface/model_factory.py", line 41, in forward
return self.model(input_ids=input_ids, decoder_input_ids=decoder_input_ids)
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 881 ops
cuda train hf_T5_base WARNING:common:fp64 golden ref were not generated for hf_T5_base. Setting accuracy check to cosine
ERROR:common:compile_fx raised LoweringException: RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s1
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s1 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
target: aten.arange.default
args[0]: s1
kwargs: {'dtype': torch.int64, 'device': device(type='cuda', index=0), 'pin_memory': False}
While executing %arange : [#users=2] = call_function[target=torch.ops.aten.arange.default](args = (%sym_size,), kwargs = {dtype: torch.int64, device: cuda:0, pin_memory: False})
Original traceback:
Module stack: {'self_model': "<class 'transformers.models.t5.modeling_t5.T5ForConditionalGeneration'>", 'self_model_encoder': "<class 'transformers.models.t5.modeling_t5.T5Stack'>", 'self_model_encoder_block_0': "<class 'transformers.models.t5.modeling_t5.T5Block'>", 'sub0_0': "<class 'transformers.models.t5.modeling_t5.T5LayerSelfAttention'>", 'self_model_encoder_block_0_layer_0_SelfAttention': "<class 'transformers.models.t5.modeling_t5.T5Attention'>"}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 423, in compute_bias
context_position = torch.arange(query_length, dtype=torch.long, device=device)[:, None]
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 519, in forward
position_bias = self.compute_bias(real_seq_length, key_length, device=scores.device)
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 570, in forward
attention_output = self.SelfAttention(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 664, in forward
self_attention_outputs = self.layer[0](
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1033, in forward
layer_outputs = layer_module(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1601, in forward
encoder_outputs = self.encoder(
| File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/util/framework/huggingface/model_factory.py", line 41, in forward
return self.model(input_ids=input_ids, decoder_input_ids=decoder_input_ids)
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 272, in call_function
out = lowerings[target](*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 221, in wrapped
return decomp_fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 1318, in arange
return fallback_arange(
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 991, in handler
TensorBox.create, ir.FallbackKernel.create(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 3011, in create
) = cls.process_kernel(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 2406, in process_kernel
example_output = kernel(*new_args, **new_kwargs)
File "/data/users/ezyang/a/pytorch/torch/_ops.py", line 500, in __call__
return self._op(*args, **kwargs or {})
File "/data/users/ezyang/a/pytorch/torch/_inductor/overrides.py", line 37, in __torch_function__
return func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_ops.py", line 500, in __call__
return self._op(*args, **kwargs or {})
RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s1
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s1 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 275, in call_function
raise LoweringException(e, target, args, kwargs) from e
torch._inductor.exc.LoweringException: RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s1
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s1 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
target: aten.arange.default
args[0]: s1
kwargs: {'dtype': torch.int64, 'device': device(type='cuda', index=0), 'pin_memory': False}
While executing %arange : [#users=2] = call_function[target=torch.ops.aten.arange.default](args = (%sym_size,), kwargs = {dtype: torch.int64, device: cuda:0, pin_memory: False})
Original traceback:
Module stack: {'self_model': "<class 'transformers.models.t5.modeling_t5.T5ForConditionalGeneration'>", 'self_model_encoder': "<class 'transformers.models.t5.modeling_t5.T5Stack'>", 'self_model_encoder_block_0': "<class 'transformers.models.t5.modeling_t5.T5Block'>", 'sub0_0': "<class 'transformers.models.t5.modeling_t5.T5LayerSelfAttention'>", 'self_model_encoder_block_0_layer_0_SelfAttention': "<class 'transformers.models.t5.modeling_t5.T5Attention'>"}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 423, in compute_bias
context_position = torch.arange(query_length, dtype=torch.long, device=device)[:, None]
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 519, in forward
position_bias = self.compute_bias(real_seq_length, key_length, device=scores.device)
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 570, in forward
attention_output = self.SelfAttention(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 664, in forward
self_attention_outputs = self.layer[0](
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1033, in forward
layer_outputs = layer_module(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1601, in forward
encoder_outputs = self.encoder(
| File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/util/framework/huggingface/model_factory.py", line 41, in forward
return self.model(input_ids=input_ids, decoder_input_ids=decoder_input_ids)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised LoweringException: RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s1
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s1 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
target: aten.arange.default
args[0]: s1
kwargs: {'dtype': torch.int64, 'device': device(type='cuda', index=0), 'pin_memory': False}
While executing %arange : [#users=2] = call_function[target=torch.ops.aten.arange.default](args = (%sym_size,), kwargs = {dtype: torch.int64, device: cuda:0, pin_memory: False})
Original traceback:
Module stack: {'self_model': "<class 'transformers.models.t5.modeling_t5.T5ForConditionalGeneration'>", 'self_model_encoder': "<class 'transformers.models.t5.modeling_t5.T5Stack'>", 'self_model_encoder_block_0': "<class 'transformers.models.t5.modeling_t5.T5Block'>", 'sub0_0': "<class 'transformers.models.t5.modeling_t5.T5LayerSelfAttention'>", 'self_model_encoder_block_0_layer_0_SelfAttention': "<class 'transformers.models.t5.modeling_t5.T5Attention'>"}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 423, in compute_bias
context_position = torch.arange(query_length, dtype=torch.long, device=device)[:, None]
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 519, in forward
position_bias = self.compute_bias(real_seq_length, key_length, device=scores.device)
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 570, in forward
attention_output = self.SelfAttention(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 664, in forward
self_attention_outputs = self.layer[0](
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1033, in forward
layer_outputs = layer_module(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1601, in forward
encoder_outputs = self.encoder(
| File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/util/framework/huggingface/model_factory.py", line 41, in forward
return self.model(input_ids=input_ids, decoder_input_ids=decoder_input_ids)
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 1643 ops
cuda train hf_T5_large PASS
Dynamo produced 0 graph(s) covering 0 ops
cuda train lennard_jones ERROR:common:'ShapeEnv' object has no attribute 'create_symbolic_sizes_strides'
While executing %tangents_1 : [#users=3] = placeholder[target=tangents_1]
Original traceback:
None
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 361, in <graph break in forward_and_backward_pass>
self.grad_scaler.scale(loss).backward()
File "/data/users/ezyang/a/pytorch/torch/_tensor.py", line 484, in backward
torch.autograd.backward(
File "/data/users/ezyang/a/pytorch/torch/autograd/__init__.py", line 197, in backward
Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 272, in apply
return user_fn(self, *args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1683, in backward
CompiledFunction.compiled_bw = aot_config.bw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 68, in _wrapped_bw_compiler
return eval_frame.disable(eval_frame.disable(bw_compiler)(*args, **kwargs))
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 380, in bw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 233, in placeholder
sizes, strides = self.symbolic_sizes_strides(example)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 50, in symbolic_sizes_strides
size, stride = self._shape_env.create_symbolic_sizes_strides(ex)
AttributeError: 'ShapeEnv' object has no attribute 'create_symbolic_sizes_strides'
While executing %tangents_1 : [#users=3] = placeholder[target=tangents_1]
Original traceback:
None
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 9 ops
cuda train maml_omniglot ERROR:common:name 's0' is not defined
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/container.py", line 202, in forward
def forward(self, input):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2340, in forward
return compiled_fn(full_args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 887, in g
return f(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1903, in debug_compiled_function
return compiled_function(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1718, in compiled_function
all_outs = CompiledFunction.apply(*args_with_synthetic_bases)
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 419, in apply
return super().apply(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1581, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 912, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 199, in run
return model(new_inputs)
File "/tmp/torchinductor_ezyang/6y/c6y6azkuvza4smupdfxq5ea47bfzgx7beluww2iqslxfm7ndhouu.py", line 330, in call
return (buf2, buf3, buf8, buf9, buf14, buf15, buf18, primals_1, primals_3, primals_5, primals_7, primals_9, primals_11, primals_24, buf1, buf2, buf3, buf4, buf5, buf19, buf7, buf8, buf9, buf10, buf11, buf20, buf13, buf14, buf15, buf16, buf21, as_strided(buf17, (5, 64), (64, 1)), as_strided(primals_13, (5, 64), (64, 1)), s0, 1, 1, )
NameError: name 's0' is not defined
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 14 ops
cuda train mnasnet1_0 ERROR:common:'ShapeEnv' object has no attribute 'create_symbolic_sizes_strides'
While executing %tangents_1 : [#users=0] = placeholder[target=tangents_1]
Original traceback:
None
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 361, in <graph break in forward_and_backward_pass>
self.grad_scaler.scale(loss).backward()
File "/data/users/ezyang/a/pytorch/torch/_tensor.py", line 484, in backward
torch.autograd.backward(
File "/data/users/ezyang/a/pytorch/torch/autograd/__init__.py", line 197, in backward
Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 272, in apply
return user_fn(self, *args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1683, in backward
CompiledFunction.compiled_bw = aot_config.bw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 68, in _wrapped_bw_compiler
return eval_frame.disable(eval_frame.disable(bw_compiler)(*args, **kwargs))
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 380, in bw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 233, in placeholder
sizes, strides = self.symbolic_sizes_strides(example)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 50, in symbolic_sizes_strides
size, stride = self._shape_env.create_symbolic_sizes_strides(ex)
AttributeError: 'ShapeEnv' object has no attribute 'create_symbolic_sizes_strides'
While executing %tangents_1 : [#users=0] = placeholder[target=tangents_1]
Original traceback:
None
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 152 ops
cuda train mobilenet_v2 ERROR:common:name 's0' is not defined
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/torchvision/torchvision/models/mobilenetv2.py", line 173, in forward
def forward(self, x: Tensor) -> Tensor:
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2340, in forward
return compiled_fn(full_args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 887, in g
return f(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1903, in debug_compiled_function
return compiled_function(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1718, in compiled_function
all_outs = CompiledFunction.apply(*args_with_synthetic_bases)
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 419, in apply
return super().apply(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1581, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 912, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 199, in run
return model(new_inputs)
File "/tmp/torchinductor_ezyang/in/cingril2qoagsybomwj7uvkez6oszpatzuvrmxongfcvpcdh6zd3.py", line 1783, in call
return (buf1, buf2, buf6, buf7, buf11, buf12, buf15, buf16, buf20, buf21, buf25, buf26, buf29, buf30, buf34, buf35, buf39, buf40, buf43, buf44, buf48, buf49, buf53, buf54, buf57, buf58, buf62, buf63, buf67, buf68, buf71, buf72, buf76, buf77, buf81, buf82, buf85, buf86, buf90, buf91, buf95, buf96, buf99, buf100, buf104, buf105, buf109, buf110, buf113, buf114, buf118, buf119, buf123, buf124, buf127, buf128, buf132, buf133, buf137, buf138, buf141, buf142, buf146, buf147, buf151, buf152, buf155, buf156, buf160, buf161, buf165, buf166, buf169, buf170, buf174, buf175, buf179, buf180, buf183, buf184, buf188, buf189, buf193, buf194, buf197, buf198, buf202, buf203, buf207, buf208, buf211, buf212, buf216, buf217, buf221, buf222, buf225, buf226, buf230, buf231, buf235, buf236, buf239, buf240, buf244, primals_1, primals_2, primals_4, primals_5, primals_7, primals_8, primals_10, primals_11, primals_13, primals_14, primals_16, primals_17, primals_19, primals_20, primals_22, primals_23, primals_25, primals_26, primals_28, primals_29, primals_31, primals_32, primals_34, primals_35, primals_37, primals_38, primals_40, primals_41, primals_43, primals_44, primals_46, primals_47, primals_49, primals_50, primals_52, primals_53, primals_55, primals_56, primals_58, primals_59, primals_61, primals_62, primals_64, primals_65, primals_67, primals_68, primals_70, primals_71, primals_73, primals_74, primals_76, primals_77, primals_79, primals_80, primals_82, primals_83, primals_85, primals_86, primals_88, primals_89, primals_91, primals_92, primals_94, primals_95, primals_97, primals_98, primals_100, primals_101, primals_103, primals_104, primals_106, primals_107, primals_109, primals_110, primals_112, primals_113, primals_115, primals_116, primals_118, primals_119, primals_121, primals_122, primals_124, primals_125, primals_127, primals_128, primals_130, primals_131, primals_133, primals_134, primals_136, primals_137, primals_139, primals_140, primals_142, primals_143, primals_145, primals_146, primals_148, primals_149, primals_151, primals_152, primals_154, primals_155, primals_315, buf0, buf1, buf2, buf4, buf5, buf6, buf7, buf9, buf10, buf11, buf12, buf13, buf14, buf15, buf16, buf18, buf19, buf20, buf21, buf23, buf24, buf25, buf26, buf27, buf28, buf29, buf30, buf32, buf33, buf34, buf35, buf37, buf38, buf39, buf40, buf41, buf42, buf43, buf44, buf46, buf47, buf48, buf49, buf51, buf52, buf53, buf54, buf55, buf56, buf57, buf58, buf60, buf61, buf62, buf63, buf65, buf66, buf67, buf68, buf69, buf70, buf71, buf72, buf74, buf75, buf76, buf77, buf79, buf80, buf81, buf82, buf83, buf84, buf85, buf86, buf88, buf89, buf90, buf91, buf93, buf94, buf95, buf96, buf97, buf98, buf99, buf100, buf102, buf103, buf104, buf105, buf107, buf108, buf109, buf110, buf111, buf112, buf113, buf114, buf116, buf117, buf118, buf119, buf121, buf122, buf123, buf124, buf125, buf126, buf127, buf128, buf130, buf131, buf132, buf133, buf135, buf136, buf137, buf138, buf139, buf140, buf141, buf142, buf144, buf145, buf146, buf147, buf149, buf150, buf151, buf152, buf153, buf154, buf155, buf156, buf158, buf159, buf160, buf161, buf163, buf164, buf165, buf166, buf167, buf168, buf169, buf170, buf172, buf173, buf174, buf175, buf177, buf178, buf179, buf180, buf181, buf182, buf183, buf184, buf186, buf187, buf188, buf189, buf191, buf192, buf193, buf194, buf195, buf196, buf197, buf198, buf200, buf201, buf202, buf203, buf205, buf206, buf207, buf208, buf209, buf210, buf211, buf212, buf214, buf215, buf216, buf217, buf219, buf220, buf221, buf222, buf223, buf224, buf225, buf226, buf228, buf229, buf230, buf231, buf233, buf234, buf235, buf236, buf237, buf238, buf239, buf240, as_strided(buf243, (4, 1280), (1280, 1)), as_strided(primals_157, (1000, 1280), (1280, 1)), buf245, buf246, buf247, buf248, buf249, buf250, buf251, buf252, buf253, buf254, buf255, buf256, buf257, buf258, buf259, buf260, buf261, buf262, buf263, buf264, buf265, buf266, buf267, buf268, buf269, buf270, buf271, buf272, buf273, buf274, buf275, buf276, buf277, buf278, buf279, s0, )
NameError: name 's0' is not defined
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 153 ops
cuda train mobilenet_v2_quantized_qat WARNING:common:fp64 golden ref were not generated for mobilenet_v2_quantized_qat. Setting accuracy check to cosine
[2022-12-12 06:47:40,143] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,154] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,175] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,180] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,188] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,205] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,210] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,217] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,236] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,243] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,262] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,267] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,274] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,294] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,300] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,307] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,327] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,334] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,353] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,359] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,366] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,382] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,388] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,395] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,415] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,419] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,426] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,443] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,448] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,454] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,474] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,480] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,487] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,506] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,513] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,533] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,539] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,546] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,564] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,570] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,577] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,595] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,599] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,606] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,624] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,630] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,637] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,655] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,661] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,672] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,690] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,695] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,701] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,719] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,726] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,732] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,752] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,758] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,766] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,786] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,793] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,814] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,821] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,828] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,847] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,854] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,862] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,881] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,885] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,892] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,912] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,919] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,927] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,946] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,961] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,968] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,988] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:40,993] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,001] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,022] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,030] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,037] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,058] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,066] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,074] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,094] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,100] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,107] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,128] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,136] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,145] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,164] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,173] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,180] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,202] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,209] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,231] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,239] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,257] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,276] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,285] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,293] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,314] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,319] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,326] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,348] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,355] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,364] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,384] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,392] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,399] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,420] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,426] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,434] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,455] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,463] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,470] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,493] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,501] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,510] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,532] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,541] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,563] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,584] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,592] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,615] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,624] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,635] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,655] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,664] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,671] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,694] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,704] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,714] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,735] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,745] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,753] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,775] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,781] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,789] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,813] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,822] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,832] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,853] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,863] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,870] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,895] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,903] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,946] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,956] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,967] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,971] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,976] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,980] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:47:41,987] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
ERROR:common:name 's0' is not defined
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/fx/graph_module.py", line 660, in call_wrapped
return self._wrapped_call(self, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/fx/graph_module.py", line 279, in __call__
raise e
File "/data/users/ezyang/a/pytorch/torch/fx/graph_module.py", line 269, in __call__
return super(self.cls, obj).__call__(*args, **kwargs) # type: ignore[misc]
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "<eval_with_key>.8", line 4, in forward
def forward(self, x : torch.Tensor) -> torch.Tensor:
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2340, in forward
return compiled_fn(full_args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 887, in g
return f(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1903, in debug_compiled_function
return compiled_function(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1718, in compiled_function
all_outs = CompiledFunction.apply(*args_with_synthetic_bases)
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 419, in apply
return super().apply(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1581, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 912, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 199, in run
return model(new_inputs)
File "/tmp/torchinductor_ezyang/g5/cg5x7byhqrl4rvqdtnu3bwzry6p6bljfufiqhw2xj4i3cbm3ngrs.py", line 7382, in call
return (buf9, buf10, buf7, buf8, buf26, buf27, buf21, buf22, buf19, buf20, buf46, buf47, buf44, buf45, buf63, buf64, buf58, buf59, buf56, buf57, buf83, buf84, buf81, buf82, buf101, buf102, buf95, buf96, buf93, buf94, buf113, buf114, buf111, buf112, buf131, buf132, buf125, buf126, buf123, buf124, buf151, buf152, buf149, buf150, buf168, buf169, buf163, buf164, buf161, buf162, buf188, buf189, buf186, buf187, buf206, buf207, buf200, buf201, buf198, buf199, buf218, buf219, buf216, buf217, buf236, buf237, buf230, buf231, buf228, buf229, buf256, buf257, buf254, buf255, buf273, buf274, buf268, buf269, buf266, buf267, buf293, buf294, buf291, buf292, buf310, buf311, buf305, buf306, buf303, buf304, buf322, buf323, buf320, buf321, buf334, buf335, buf332, buf333, buf351, buf352, buf346, buf347, buf344, buf345, buf371, buf372, buf369, buf370, buf388, buf389, buf383, buf384, buf381, buf382, buf408, buf409, buf406, buf407, buf425, buf426, buf420, buf421, buf418, buf419, buf437, buf438, buf435, buf436, buf455, buf456, buf449, buf450, buf447, buf448, buf475, buf476, buf473, buf474, buf492, buf493, buf487, buf488, buf485, buf486, buf512, buf513, buf510, buf511, buf529, buf530, buf524, buf525, buf522, buf523, buf541, buf542, buf539, buf540, buf553, buf554, buf551, buf552, buf570, buf571, buf565, buf566, buf563, buf564, buf590, buf591, buf588, buf589, buf607, buf608, buf602, buf603, buf600, buf601, buf627, buf628, buf625, buf626, buf644, buf645, buf639, buf640, buf637, buf638, buf656, buf657, buf654, buf655, buf668, buf669, buf666, buf667, buf685, buf686, buf680, buf681, buf678, buf679, buf705, buf706, buf703, buf704, buf722, buf723, buf717, buf718, buf715, buf716, buf742, buf743, buf740, buf741, buf760, buf761, buf754, buf755, buf752, buf753, buf772, buf773, buf770, buf771, buf790, buf791, buf784, buf785, buf782, buf783, buf810, buf811, buf808, buf809, buf827, buf828, buf822, buf823, buf820, buf821, buf847, buf848, buf845, buf846, buf864, buf865, buf859, buf860, buf857, buf858, buf876, buf877, buf874, buf875, buf888, buf889, buf886, buf887, buf905, buf906, buf900, buf901, buf898, buf899, buf925, buf926, buf923, buf924, buf942, buf943, buf937, buf938, buf935, buf936, buf962, buf963, buf960, buf961, buf979, buf980, buf974, buf975, buf972, buf973, buf991, buf992, buf989, buf990, buf1003, buf1004, buf1001, buf1002, buf1020, buf1021, buf1015, buf1016, buf1013, buf1014, buf1040, buf1041, buf1038, buf1039, buf1057, buf1058, buf1052, buf1053, buf1050, buf1051, buf1077, buf1078, buf1075, buf1076, buf1094, buf1095, buf1089, buf1090, buf1087, buf1088, buf1106, buf1107, buf1104, buf1105, buf1118, buf1119, buf1116, buf1117, buf1135, buf1136, buf1130, buf1131, buf1128, buf1129, buf1155, buf1156, buf1153, buf1154, buf1172, buf1173, buf1167, buf1168, buf1165, buf1166, buf1192, buf1193, buf1190, buf1191, buf1209, buf1210, buf1204, buf1205, buf1202, buf1203, buf1221, buf1222, buf1219, buf1220, buf1239, buf1240, buf1233, buf1234, buf1231, buf1232, buf1259, buf1260, buf1257, buf1258, buf1276, buf1277, buf1271, buf1272, buf1269, buf1270, buf1296, buf1297, buf1294, buf1295, buf1313, buf1314, buf1308, buf1309, buf1306, buf1307, buf1325, buf1326, buf1323, buf1324, buf1337, buf1338, buf1335, buf1336, buf1354, buf1355, buf1349, buf1350, buf1347, buf1348, buf1374, buf1375, buf1372, buf1373, buf1391, buf1392, buf1386, buf1387, buf1384, buf1385, buf1411, buf1412, buf1409, buf1410, buf1428, buf1429, buf1423, buf1424, buf1421, buf1422, buf1440, buf1441, buf1438, buf1439, buf1452, buf1453, buf1450, buf1451, buf1469, buf1470, buf1464, buf1465, buf1462, buf1463, buf1489, buf1490, buf1487, buf1488, buf1506, buf1507, buf1501, buf1502, buf1499, buf1500, buf1526, buf1527, buf1524, buf1525, buf1544, buf1545, buf1538, buf1539, buf1536, buf1537, buf1556, buf1557, buf1554, buf1555, buf1574, buf1575, buf1568, buf1569, buf1566, buf1567, buf1594, buf1595, buf1592, buf1593, buf1611, buf1612, buf1606, buf1607, buf1604, buf1605, buf1631, buf1632, buf1629, buf1630, buf1648, buf1649, buf1643, buf1644, buf1641, buf1642, buf1660, buf1661, buf1658, buf1659, buf1672, buf1673, buf1670, buf1671, buf1689, buf1690, buf1684, buf1685, buf1682, buf1683, buf1709, buf1710, buf1707, buf1708, buf1726, buf1727, buf1721, buf1722, buf1719, buf1720, buf1746, buf1747, buf1744, buf1745, buf1763, buf1764, buf1758, buf1759, buf1756, buf1757, buf1775, buf1776, buf1773, buf1774, buf1787, buf1788, buf1785, buf1786, buf1804, buf1805, buf1799, buf1800, buf1797, buf1798, buf1824, buf1825, buf1822, buf1823, buf1841, buf1842, buf1836, buf1837, buf1834, buf1835, buf1861, buf1862, buf1859, buf1860, buf1879, buf1880, buf1873, buf1874, buf1871, buf1872, buf1891, buf1892, buf1889, buf1890, buf1909, buf1910, buf1903, buf1904, buf1901, buf1902, buf1945, buf1946, buf1943, buf1944, buf1956, buf1957, buf1954, buf1955, buf1967, buf1968, buf1965, buf1966, buf1979, buf1980, buf1977, buf1978, buf1975, primals_1, primals_2, primals_4, primals_5, primals_7, primals_8, primals_10, primals_11, primals_13, primals_14, primals_16, primals_17, primals_19, primals_20, primals_22, primals_23, primals_25, primals_26, primals_28, primals_29, primals_31, primals_32, primals_34, primals_35, primals_37, primals_38, primals_40, primals_41, primals_43, primals_44, primals_46, primals_47, primals_49, primals_50, primals_52, primals_53, primals_55, primals_56, primals_58, primals_59, primals_61, primals_62, primals_64, primals_65, primals_67, primals_68, primals_70, primals_71, primals_73, primals_74, primals_76, primals_77, primals_79, primals_80, primals_82, primals_83, primals_85, primals_86, primals_88, primals_89, primals_91, primals_92, primals_94, primals_95, primals_97, primals_98, primals_100, primals_101, primals_103, primals_104, primals_106, primals_107, primals_109, primals_110, primals_112, primals_113, primals_115, primals_116, primals_118, primals_119, primals_121, primals_122, primals_124, primals_125, primals_127, primals_128, primals_130, primals_131, primals_133, primals_134, primals_136, primals_137, primals_139, primals_140, primals_142, primals_143, primals_145, primals_146, primals_148, primals_149, primals_151, primals_152, primals_154, primals_155, primals_167, primals_184, primals_201, primals_218, primals_235, primals_252, primals_269, primals_286, primals_303, primals_327, primals_344, primals_361, primals_378, primals_395, primals_412, primals_436, primals_453, primals_470, primals_494, primals_511, primals_528, primals_545, primals_562, primals_579, primals_603, primals_620, primals_637, primals_661, primals_678, primals_695, primals_719, primals_736, primals_753, primals_770, primals_787, primals_804, primals_828, primals_845, primals_862, primals_886, primals_903, primals_920, primals_937, primals_954, primals_971, primals_995, primals_1012, primals_1029, primals_1053, primals_1070, primals_1087, primals_1104, buf5, buf17, buf18, buf25, buf26, buf35, buf42, buf43, buf60, buf55, buf62, buf63, buf72, buf79, buf80, buf97, buf92, buf100, buf101, buf109, buf110, buf127, buf122, buf130, buf131, buf140, buf147, buf148, buf165, buf160, buf167, buf168, buf177, buf184, buf185, buf202, buf197, buf205, buf206, buf214, buf215, buf232, buf227, buf235, buf236, buf245, buf252, buf253, buf270, buf265, buf272, buf273, buf282, buf289, buf290, buf307, buf302, buf309, buf310, buf319, buf330, buf331, buf348, buf343, buf350, buf351, buf360, buf367, buf368, buf385, buf380, buf387, buf388, buf397, buf404, buf405, buf422, buf417, buf424, buf425, buf433, buf434, buf451, buf446, buf454, buf455, buf464, buf471, buf472, buf489, buf484, buf491, buf492, buf501, buf508, buf509, buf526, buf521, buf528, buf529, buf538, buf549, buf550, buf567, buf562, buf569, buf570, buf579, buf586, buf587, buf604, buf599, buf606, buf607, buf616, buf623, buf624, buf641, buf636, buf643, buf644, buf653, buf664, buf665, buf682, buf677, buf684, buf685, buf694, buf701, buf702, buf719, buf714, buf721, buf722, buf731, buf738, buf739, buf756, buf751, buf759, buf760, buf768, buf769, buf786, buf781, buf789, buf790, buf799, buf806, buf807, buf824, buf819, buf826, buf827, buf836, buf843, buf844, buf861, buf856, buf863, buf864, buf873, buf884, buf885, buf902, buf897, buf904, buf905, buf914, buf921, buf922, buf939, buf934, buf941, buf942, buf951, buf958, buf959, buf976, buf971, buf978, buf979, buf988, buf999, buf1000, buf1017, buf1012, buf1019, buf1020, buf1029, buf1036, buf1037, buf1054, buf1049, buf1056, buf1057, buf1066, buf1073, buf1074, buf1091, buf1086, buf1093, buf1094, buf1103, buf1114, buf1115, buf1132, buf1127, buf1134, buf1135, buf1144, buf1151, buf1152, buf1169, buf1164, buf1171, buf1172, buf1181, buf1188, buf1189, buf1206, buf1201, buf1208, buf1209, buf1217, buf1218, buf1235, buf1230, buf1238, buf1239, buf1248, buf1255, buf1256, buf1273, buf1268, buf1275, buf1276, buf1285, buf1292, buf1293, buf1310, buf1305, buf1312, buf1313, buf1322, buf1333, buf1334, buf1351, buf1346, buf1353, buf1354, buf1363, buf1370, buf1371, buf1388, buf1383, buf1390, buf1391, buf1400, buf1407, buf1408, buf1425, buf1420, buf1427, buf1428, buf1437, buf1448, buf1449, buf1466, buf1461, buf1468, buf1469, buf1478, buf1485, buf1486, buf1503, buf1498, buf1505, buf1506, buf1515, buf1522, buf1523, buf1540, buf1535, buf1543, buf1544, buf1552, buf1553, buf1570, buf1565, buf1573, buf1574, buf1583, buf1590, buf1591, buf1608, buf1603, buf1610, buf1611, buf1620, buf1627, buf1628, buf1645, buf1640, buf1647, buf1648, buf1657, buf1668, buf1669, buf1686, buf1681, buf1688, buf1689, buf1698, buf1705, buf1706, buf1723, buf1718, buf1725, buf1726, buf1735, buf1742, buf1743, buf1760, buf1755, buf1762, buf1763, buf1772, buf1783, buf1784, buf1801, buf1796, buf1803, buf1804, buf1813, buf1820, buf1821, buf1838, buf1833, buf1840, buf1841, buf1850, buf1857, buf1858, buf1875, buf1870, buf1878, buf1879, buf1887, buf1888, buf1905, buf1900, buf1908, buf1909, buf1918, buf1926, buf1935, buf1942, buf1952, buf1953, buf1964, buf1976, as_strided(buf1963, (1000, 1280), (1280, 1)), buf1981, buf1982, buf1983, buf1984, buf1985, buf1986, buf1987, buf1988, buf1989, buf1990, buf1991, buf1992, buf1993, buf1994, buf1995, buf1996, buf1997, buf1998, buf1999, buf2000, buf2001, buf2002, buf2003, buf2004, buf2005, buf2006, buf2007, buf2008, buf2009, buf2010, buf2011, buf2012, buf2013, buf2014, buf2015, s0, )
NameError: name 's0' is not defined
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 203 ops
cuda train mobilenet_v3_large ERROR:common:name 's0' is not defined
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/torchvision/torchvision/models/mobilenetv3.py", line 219, in forward
def forward(self, x: Tensor) -> Tensor:
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2340, in forward
return compiled_fn(full_args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 887, in g
return f(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1903, in debug_compiled_function
return compiled_function(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1718, in compiled_function
all_outs = CompiledFunction.apply(*args_with_synthetic_bases)
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 419, in apply
return super().apply(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1581, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 912, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 199, in run
return model(new_inputs)
File "/tmp/torchinductor_ezyang/bb/cbb6eyb435nax2dw5el7ux4pqvdo3hiiovjwiopdwipnb752ak45.py", line 2733, in call
return (buf1, buf2, buf6, buf7, buf10, buf11, buf14, buf15, buf18, buf19, buf22, buf23, buf26, buf27, buf30, buf31, buf34, buf35, buf38, buf39, buf42, buf43, buf53, buf54, buf57, buf58, buf61, buf62, buf72, buf73, buf76, buf77, buf80, buf81, buf91, buf92, buf95, buf96, buf100, buf101, buf105, buf106, buf109, buf110, buf114, buf115, buf119, buf120, buf123, buf124, buf128, buf129, buf133, buf134, buf137, buf138, buf142, buf143, buf147, buf148, buf151, buf152, buf156, buf157, buf168, buf169, buf172, buf173, buf177, buf178, buf189, buf190, buf193, buf194, buf198, buf199, buf210, buf211, buf214, buf215, buf219, buf220, buf231, buf232, buf235, buf236, buf240, buf241, buf252, buf253, buf256, buf257, buf263, primals_1, primals_2, primals_4, primals_5, primals_7, primals_8, primals_10, primals_11, primals_13, primals_14, primals_16, primals_17, primals_19, primals_20, primals_22, primals_23, primals_25, primals_26, primals_28, primals_29, primals_31, primals_32, primals_34, primals_36, primals_38, primals_39, primals_41, primals_42, primals_44, primals_45, primals_47, primals_49, primals_51, primals_52, primals_54, primals_55, primals_57, primals_58, primals_60, primals_62, primals_64, primals_65, primals_67, primals_68, primals_70, primals_71, primals_73, primals_74, primals_76, primals_77, primals_79, primals_80, primals_82, primals_83, primals_85, primals_86, primals_88, primals_89, primals_91, primals_92, primals_94, primals_95, primals_97, primals_98, primals_100, primals_101, primals_103, primals_104, primals_106, primals_107, primals_109, primals_111, primals_113, primals_114, primals_116, primals_117, primals_119, primals_120, primals_122, primals_124, primals_126, primals_127, primals_129, primals_130, primals_132, primals_133, primals_135, primals_137, primals_139, primals_140, primals_142, primals_143, primals_145, primals_146, primals_148, primals_150, primals_152, primals_153, primals_155, primals_156, primals_158, primals_159, primals_161, primals_163, primals_165, primals_166, primals_168, primals_169, primals_313, buf0, buf1, buf2, buf264, buf4, buf5, buf6, buf7, buf8, buf9, buf10, buf11, buf12, buf13, buf14, buf15, buf16, buf17, buf18, buf19, buf20, buf21, buf22, buf23, buf24, buf25, buf26, buf27, buf28, buf29, buf30, buf31, buf32, buf33, buf34, buf35, buf36, buf37, buf38, buf39, buf40, buf41, buf42, buf43, buf44, buf46, buf48, buf50, buf52, buf53, buf54, buf55, buf56, buf57, buf58, buf59, buf60, buf61, buf62, buf63, buf65, buf67, buf69, buf71, buf72, buf73, buf74, buf75, buf76, buf77, buf78, buf79, buf80, buf81, buf82, buf84, buf86, buf88, buf90, buf91, buf92, buf93, buf94, buf95, buf96, buf265, buf98, buf99, buf100, buf101, buf266, buf103, buf104, buf105, buf106, buf107, buf108, buf109, buf110, buf267, buf112, buf113, buf114, buf115, buf268, buf117, buf118, buf119, buf120, buf121, buf122, buf123, buf124, buf269, buf126, buf127, buf128, buf129, buf270, buf131, buf132, buf133, buf134, buf135, buf136, buf137, buf138, buf271, buf140, buf141, buf142, buf143, buf272, buf145, buf146, buf147, buf148, buf149, buf150, buf151, buf152, buf273, buf154, buf155, buf156, buf157, buf274, buf159, buf161, buf163, buf165, buf167, buf168, buf169, buf170, buf171, buf172, buf173, buf275, buf175, buf176, buf177, buf178, buf276, buf180, buf182, buf184, buf186, buf188, buf189, buf190, buf191, buf192, buf193, buf194, buf277, buf196, buf197, buf198, buf199, buf278, buf201, buf203, buf205, buf207, buf209, buf210, buf211, buf212, buf213, buf214, buf215, buf279, buf217, buf218, buf219, buf220, buf280, buf222, buf224, buf226, buf228, buf230, buf231, buf232, buf233, buf234, buf235, buf236, buf281, buf238, buf239, buf240, buf241, buf282, buf243, buf245, buf247, buf249, buf251, buf252, buf253, buf254, buf255, buf256, buf257, buf283, as_strided(buf260, (4, 960), (960, 1)), buf261, buf262, as_strided(primals_173, (1000, 1280), (1280, 1)), as_strided(primals_171, (1280, 960), (960, 1)), buf284, buf285, buf286, buf287, buf288, buf289, buf290, buf291, s0, )
NameError: name 's0' is not defined
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 187 ops
devgpu019:3716426:3716426 [0] NCCL INFO NCCL_SOCKET_IFNAME set by environment to eth0
devgpu019:3716426:3716426 [0] NCCL INFO NCCL_SOCKET_IFNAME set to eth0
devgpu019:3716426:3716426 [0] NCCL INFO Bootstrap : Using eth0:2803:6080:6188:70b4::1<0>
devgpu019:3716426:3716426 [0] NCCL INFO NET/Plugin : No plugin found (libnccl-net.so), using internal implementation
devgpu019:3716426:3716426 [0] NCCL INFO cudaDriverVersion 11040
NCCL version 2.14.3+cuda11.4
devgpu019:3716426:3718755 [0] NCCL INFO NCCL_IB_DISABLE set by environment to 1.
devgpu019:3716426:3718755 [0] NCCL INFO NCCL_SOCKET_IFNAME set by environment to eth0
devgpu019:3716426:3718755 [0] NCCL INFO NET/Socket : Using [0]eth0:2803:6080:6188:70b4::1<0>
devgpu019:3716426:3718755 [0] NCCL INFO Using network Socket
devgpu019:3716426:3718755 [0] NCCL INFO NET/Socket : GPU Direct RDMA Disabled for HCA 0 'eth0'
devgpu019:3716426:3718755 [0] NCCL INFO === System : maxBw 5000.0 totalBw 0.0 ===
devgpu019:3716426:3718755 [0] NCCL INFO CPU/0 (1/1/2)
devgpu019:3716426:3718755 [0] NCCL INFO + PCI[12.0] - PCI/D000 (11f840001d9bfbe1)
devgpu019:3716426:3718755 [0] NCCL INFO + PCI[24.0] - PCI/F000 (11f840001d9bfbe0)
devgpu019:3716426:3718755 [0] NCCL INFO + PCI[24.0] - GPU/11000 (0)
devgpu019:3716426:3718755 [0] NCCL INFO + PCI[12.0] - NIC/30000
devgpu019:3716426:3718755 [0] NCCL INFO ==========================================
devgpu019:3716426:3718755 [0] NCCL INFO GPU/11000 :GPU/11000 (0/5000.000000/LOC) CPU/0 (3/12.000000/PHB)
devgpu019:3716426:3718755 [0] NCCL INFO Setting affinity for GPU 0 to ffffff,00000000,00000000,00ffffff
devgpu019:3716426:3718755 [0] NCCL INFO Pattern 4, crossNic 0, nChannels 16, bw 44.000000/44.000000, type LOC/PIX, sameChannels 1
devgpu019:3716426:3718755 [0] NCCL INFO 0 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 1 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 2 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 3 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 4 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 5 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 6 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 7 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 8 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 9 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 10 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 11 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 12 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 13 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 14 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 15 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO Pattern 3, crossNic 0, nChannels 16, bw 44.000000/44.000000, type LOC/PIX, sameChannels 1
devgpu019:3716426:3718755 [0] NCCL INFO 0 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 1 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 2 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 3 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 4 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 5 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 6 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 7 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 8 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 9 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 10 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 11 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 12 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 13 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 14 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 15 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO Pattern 3, crossNic 0, nChannels 16, bw 44.000000/44.000000, type LOC/PIX, sameChannels 1
devgpu019:3716426:3718755 [0] NCCL INFO 0 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 1 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 2 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 3 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 4 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 5 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 6 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 7 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 8 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 9 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 10 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 11 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 12 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 13 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 14 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO 15 : GPU/0
devgpu019:3716426:3718755 [0] NCCL INFO Tree 0 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 16 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 1 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 17 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 2 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 18 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 3 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 19 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 4 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 20 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 5 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 21 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 6 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 22 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 7 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 23 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 8 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 24 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 9 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 25 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 10 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 26 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 11 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 27 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 12 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 28 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 13 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 29 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 14 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 30 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 15 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Tree 31 : -1 -> 0 -> -1/-1/-1
devgpu019:3716426:3718755 [0] NCCL INFO Channel 00/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 01/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 02/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 03/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 04/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 05/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 06/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 07/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 08/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 09/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 10/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 11/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 12/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 13/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 14/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 15/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 16/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 17/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 18/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 19/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 20/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 21/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 22/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 23/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 24/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 25/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 26/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 27/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 28/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 29/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 30/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Channel 31/32 : 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 00 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 01 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 02 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 03 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 04 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 05 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 06 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 07 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 08 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 09 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 10 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 11 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 12 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 13 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 14 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 15 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 16 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 17 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 18 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 19 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 20 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 21 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 22 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 23 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 24 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 25 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 26 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 27 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 28 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 29 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 30 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Ring 31 : 0 -> 0 -> 0
devgpu019:3716426:3718755 [0] NCCL INFO Trees [0] -1/-1/-1->0->-1 [1] -1/-1/-1->0->-1 [2] -1/-1/-1->0->-1 [3] -1/-1/-1->0->-1 [4] -1/-1/-1->0->-1 [5] -1/-1/-1->0->-1 [6] -1/-1/-1->0->-1 [7] -1/-1/-1->0->-1 [8] -1/-1/-1->0->-1 [9] -1/-1/-1->0->-1 [10] -1/-1/-1->0->-1 [11] -1/-1/-1->0->-1 [12] -1/-1/-1->0->-1 [13] -1/-1/-1->0->-1 [14] -1/-1/-1->0->-1 [15] -1/-1/-1->0->-1 [16] -1/-1/-1->0->-1 [17] -1/-1/-1->0->-1 [18] -1/-1/-1->0->-1 [19] -1/-1/-1->0->-1 [20] -1/-1/-1->0->-1 [21] -1/-1/-1->0->-1 [22] -1/-1/-1->0->-1 [23] -1/-1/-1->0->-1 [24] -1/-1/-1->0->-1 [25] -1/-1/-1->0->-1 [26] -1/-1/-1->0->-1 [27] -1/-1/-1->0->-1 [28] -1/-1/-1->0->-1 [29] -1/-1/-1->0->-1 [30] -1/-1/-1->0->-1 [31] -1/-1/-1->0->-1
devgpu019:3716426:3718755 [0] NCCL INFO Connected all rings
devgpu019:3716426:3718755 [0] NCCL INFO Connected all trees
devgpu019:3716426:3718755 [0] NCCL INFO 32 coll channels, 32 p2p channels, 32 p2p channels per peer
devgpu019:3716426:3718772 [0] NCCL INFO New proxy send connection 0 from local rank 0, transport 2
devgpu019:3716426:3718755 [0] NCCL INFO Connection to proxy localRank 0 -> connection 0x7fbf58002e80
devgpu019:3716426:3718755 [0] NCCL INFO comm 0x71c1fc0 rank 0 nranks 1 cudaDev 0 busId 11000 - Init COMPLETE
cuda train moco [2022-12-12 06:49:45,255] torch._dynamo.convert_frame: [WARNING] torch._dynamo hit config.cache_size_limit (64)
function: '<graph break in _momentum_update_key_encoder>' (/data/users/ezyang/a/torchbenchmark/torchbenchmark/models/moco/moco/builder.py:50)
reasons: ___tuple_iterator_len(___stack0) == 160
to diagnose recompilation issues, see https://github.com/pytorch/torchdynamo/blob/main/TROUBLESHOOTING.md.
ERROR:common:
from user code:
File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/models/moco/moco/builder.py", line 172, in concat_all_gather
torch.distributed.all_gather(tensors_gather, tensor, async_op=False)
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 1087, in run_node
return node.target(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/distributed/distributed_c10d.py", line 1346, in wrapper
return func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/distributed/distributed_c10d.py", line 2341, in all_gather
work = default_pg.allgather([tensor_list], [tensor])
File "/data/users/ezyang/a/pytorch/torch/_subclasses/fake_tensor.py", line 896, in __torch_dispatch__
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_ops.py", line 285, in __call__
return self._op(*args, **kwargs or {})
RuntimeError: Tensors must be CUDA and dense
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 1046, in get_fake_value
return wrap_fake_exception(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 712, in wrap_fake_exception
return fn()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 1047, in <lambda>
lambda: run_node(tx.output, node, args, kwargs, nnmodule)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 1096, in run_node
raise RuntimeError(
RuntimeError: Failed running call_function <function all_gather at 0x7fc85908dee0>(*([FakeTensor(FakeTensor(..., device='meta', size=(s0, s1, s2, s2)), cuda:0)], FakeTensor(FakeTensor(..., device='meta', size=(s0, s1, s2, s2)), cuda:0)), **{'async_op': False}):
Tensors must be CUDA and dense
(scroll up for backtrace)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/nn/parallel/distributed.py", line 1098, in forward
output = self._run_ddp_forward(*inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/nn/parallel/distributed.py", line 1051, in _run_ddp_forward
return module_to_run(*inputs[0], **kwargs[0]) # type: ignore[index]
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/models/moco/moco/builder.py", line 130, in forward
self._momentum_update_key_encoder() # update the key encoder
File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/models/moco/moco/builder.py", line 133, in <graph break in forward>
im_k, idx_unshuffle = self._batch_shuffle_ddp(im_k)
File "/data/users/ezyang/a/pytorch/torch/autograd/grad_mode.py", line 34, in decorate_context
return func(*args, **kwargs)
File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/models/moco/moco/builder.py", line 76, in _batch_shuffle_ddp
x_gather = concat_all_gather(x)
File "/data/users/ezyang/a/pytorch/torch/autograd/grad_mode.py", line 34, in decorate_context
return func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 329, in catch_errors
return hijacked_callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 305, in wrapper
return inner_fn(self, inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1005, in CALL_FUNCTION_KW
self.call_function(fn, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 430, in call_function
self.push(fn.call_function(self, args, kwargs))
File "/data/users/ezyang/a/pytorch/torch/_dynamo/variables/torch.py", line 444, in call_function
tensor_variable = wrap_fx_proxy(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/variables/builder.py", line 729, in wrap_fx_proxy
return wrap_fx_proxy_cls(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/variables/builder.py", line 766, in wrap_fx_proxy_cls
example_value = get_fake_value(proxy.node, tx)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 1066, in get_fake_value
raise TorchRuntimeError() from e
torch._dynamo.exc.TorchRuntimeError:
from user code:
File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/models/moco/moco/builder.py", line 172, in concat_all_gather
torch.distributed.all_gather(tensors_gather, tensor, async_op=False)
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 68 graph(s) covering 507 ops
devgpu019:3716426:3718772 [0] NCCL INFO [Service thread] Connection closed by localRank 0
devgpu019:3716426:3716426 [0] NCCL INFO comm 0x71c1fc0 rank 0 nranks 1 cudaDev 0 busId 11000 - Abort COMPLETE
cuda train nvidia_deeprecommender ERROR:common:'ShapeEnv' object has no attribute 'create_symbolic_sizes_strides'
While executing %tangents_1 : [#users=2] = placeholder[target=tangents_1]
Original traceback:
None
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 361, in <graph break in forward_and_backward_pass>
self.grad_scaler.scale(loss).backward()
File "/data/users/ezyang/a/pytorch/torch/_tensor.py", line 484, in backward
torch.autograd.backward(
File "/data/users/ezyang/a/pytorch/torch/autograd/__init__.py", line 197, in backward
Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 272, in apply
return user_fn(self, *args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1683, in backward
CompiledFunction.compiled_bw = aot_config.bw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 68, in _wrapped_bw_compiler
return eval_frame.disable(eval_frame.disable(bw_compiler)(*args, **kwargs))
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 380, in bw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 233, in placeholder
sizes, strides = self.symbolic_sizes_strides(example)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 50, in symbolic_sizes_strides
size, stride = self._shape_env.create_symbolic_sizes_strides(ex)
AttributeError: 'ShapeEnv' object has no attribute 'create_symbolic_sizes_strides'
While executing %tangents_1 : [#users=2] = placeholder[target=tangents_1]
Original traceback:
None
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 13 ops
--dataroot /data/users/ezyang/a/torchbenchmark/torchbenchmark/data/.data/pytorch_CycleGAN_and_pix2pix_inputs/datasets/horse2zebra --name horse2zebra --model cycle_gan --display_id 0 --n_epochs 3 --n_epochs_decay 3 --gpu_ids 0 --checkpoints_dir /data/users/ezyang/a/torchbenchmark/torchbenchmark/models/pytorch_CycleGAN_and_pix2pix/.data/checkpoints
cuda train pytorch_CycleGAN_and_pix2pix ERROR:common:'int' object has no attribute 'size'
While executing %sym_size : [#users=74] = placeholder[target=sym_size]
Original traceback:
Module stack: {'self_model': "<class 'torch.nn.modules.container.Sequential'>", 'self_model_2': "<class 'torch.nn.modules.instancenorm.InstanceNorm2d'>"}
File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/models/pytorch_CycleGAN_and_pix2pix/models/networks.py", line 372, in forward
return self.model(input)
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 361, in <graph break in forward_and_backward_pass>
self.grad_scaler.scale(loss).backward()
File "/data/users/ezyang/a/pytorch/torch/_tensor.py", line 484, in backward
torch.autograd.backward(
File "/data/users/ezyang/a/pytorch/torch/autograd/__init__.py", line 197, in backward
Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 272, in apply
return user_fn(self, *args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1683, in backward
CompiledFunction.compiled_bw = aot_config.bw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 68, in _wrapped_bw_compiler
return eval_frame.disable(eval_frame.disable(bw_compiler)(*args, **kwargs))
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 380, in bw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 231, in placeholder
sizes, strides = self.static_sizes_strides(example)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 62, in static_sizes_strides
size = [sympy.Integer(i) for i in ex.size()]
AttributeError: 'int' object has no attribute 'size'
While executing %sym_size : [#users=74] = placeholder[target=sym_size]
Original traceback:
Module stack: {'self_model': "<class 'torch.nn.modules.container.Sequential'>", 'self_model_2': "<class 'torch.nn.modules.instancenorm.InstanceNorm2d'>"}
File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/models/pytorch_CycleGAN_and_pix2pix/models/networks.py", line 372, in forward
return self.model(input)
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 91 ops
cuda train pytorch_stargan ERROR:common:compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 440, in codegen
self.scheduler.codegen()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/scheduler.py", line 1129, in codegen
self.get_backend(device).codegen_nodes(node.get_nodes())
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1284, in codegen_nodes
return self.codegen_node_schedule(node_schedule, numel, rnumel)
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1336, in codegen_node_schedule
src_code = kernel.codegen_kernel()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1035, in codegen_kernel
"signature": dict(enumerate(map(signature_of, signature))),
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 57, in signature_of
return JITFunction._key_of(V.graph.sizevars.size_hint(arg.expr))
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 367, in size_hint
return int(out)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 476, in compile_subgraph
self.compile_and_call_fx_graph(tx, list(reversed(stack_values)), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 60 ops
cuda train pytorch_struct ERROR:common:compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 440, in codegen
self.scheduler.codegen()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/scheduler.py", line 1129, in codegen
self.get_backend(device).codegen_nodes(node.get_nodes())
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1284, in codegen_nodes
return self.codegen_node_schedule(node_schedule, numel, rnumel)
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1336, in codegen_node_schedule
src_code = kernel.codegen_kernel()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1035, in codegen_kernel
"signature": dict(enumerate(map(signature_of, signature))),
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 57, in signature_of
return JITFunction._key_of(V.graph.sizevars.size_hint(arg.expr))
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 367, in size_hint
return int(out)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 47 ops
cuda train pytorch_unet [2022-12-12 06:51:55,058] torch._inductor.graph: [WARNING] Creating implicit fallback for:
target: <function sym_float at 0x7fdd4f725ca0>
args[0]: 80.0
ERROR:common:compile_fx raised LoweringException: TypeError: sym_float() missing 1 required positional argument: 'a'
target: <function sym_float at 0x7fdd4f725ca0>
args[0]: 80.0
While executing %sym_float : [#users=1] = call_function[target=torch.fx.experimental.symbolic_shapes.sym_float](args = (%mul_30,), kwargs = {})
Original traceback:
None
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 272, in call_function
out = lowerings[target](*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 221, in wrapped
return decomp_fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 991, in handler
TensorBox.create, ir.FallbackKernel.create(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 3011, in create
) = cls.process_kernel(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 2406, in process_kernel
example_output = kernel(*new_args, **new_kwargs)
TypeError: sym_float() missing 1 required positional argument: 'a'
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 275, in call_function
raise LoweringException(e, target, args, kwargs) from e
torch._inductor.exc.LoweringException: TypeError: sym_float() missing 1 required positional argument: 'a'
target: <function sym_float at 0x7fdd4f725ca0>
args[0]: 80.0
While executing %sym_float : [#users=1] = call_function[target=torch.fx.experimental.symbolic_shapes.sym_float](args = (%mul_30,), kwargs = {})
Original traceback:
None
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 476, in compile_subgraph
self.compile_and_call_fx_graph(tx, list(reversed(stack_values)), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised LoweringException: TypeError: sym_float() missing 1 required positional argument: 'a'
target: <function sym_float at 0x7fdd4f725ca0>
args[0]: 80.0
While executing %sym_float : [#users=1] = call_function[target=torch.fx.experimental.symbolic_shapes.sym_float](args = (%mul_30,), kwargs = {})
Original traceback:
None
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 135 ops
cuda train resnet152 ERROR:common:name 's0' is not defined
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/torchvision/torchvision/models/resnet.py", line 284, in forward
def forward(self, x: Tensor) -> Tensor:
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2340, in forward
return compiled_fn(full_args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 887, in g
return f(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1903, in debug_compiled_function
return compiled_function(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1718, in compiled_function
all_outs = CompiledFunction.apply(*args_with_synthetic_bases)
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 419, in apply
return super().apply(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1581, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 912, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 199, in run
return model(new_inputs)
File "/tmp/torchinductor_ezyang/52/c52pgv3xeogyipebdwld3wxarzy7mb4c2fu7hartsogr4dmvjvzs.py", line 2691, in call
return (buf1, buf2, buf7, buf8, buf11, buf12, buf15, buf16, buf18, buf19, buf23, buf24, buf27, buf28, buf31, buf32, buf35, buf36, buf39, buf40, buf43, buf44, buf47, buf48, buf51, buf52, buf55, buf56, buf58, buf59, buf63, buf64, buf67, buf68, buf71, buf72, buf75, buf76, buf79, buf80, buf83, buf84, buf87, buf88, buf91, buf92, buf95, buf96, buf99, buf100, buf103, buf104, buf107, buf108, buf111, buf112, buf115, buf116, buf119, buf120, buf123, buf124, buf127, buf128, buf131, buf132, buf135, buf136, buf139, buf140, buf143, buf144, buf147, buf148, buf151, buf152, buf155, buf156, buf158, buf159, buf163, buf164, buf167, buf168, buf171, buf172, buf175, buf176, buf179, buf180, buf183, buf184, buf187, buf188, buf191, buf192, buf195, buf196, buf199, buf200, buf203, buf204, buf207, buf208, buf211, buf212, buf215, buf216, buf219, buf220, buf223, buf224, buf227, buf228, buf231, buf232, buf235, buf236, buf239, buf240, buf243, buf244, buf247, buf248, buf251, buf252, buf255, buf256, buf259, buf260, buf263, buf264, buf267, buf268, buf271, buf272, buf275, buf276, buf279, buf280, buf283, buf284, buf287, buf288, buf291, buf292, buf295, buf296, buf299, buf300, buf303, buf304, buf307, buf308, buf311, buf312, buf315, buf316, buf319, buf320, buf323, buf324, buf327, buf328, buf331, buf332, buf335, buf336, buf339, buf340, buf343, buf344, buf347, buf348, buf351, buf352, buf355, buf356, buf359, buf360, buf363, buf364, buf367, buf368, buf371, buf372, buf375, buf376, buf379, buf380, buf383, buf384, buf387, buf388, buf391, buf392, buf395, buf396, buf399, buf400, buf403, buf404, buf407, buf408, buf411, buf412, buf415, buf416, buf419, buf420, buf423, buf424, buf427, buf428, buf431, buf432, buf435, buf436, buf439, buf440, buf443, buf444, buf447, buf448, buf451, buf452, buf455, buf456, buf459, buf460, buf463, buf464, buf467, buf468, buf471, buf472, buf475, buf476, buf479, buf480, buf483, buf484, buf487, buf488, buf491, buf492, buf495, buf496, buf499, buf500, buf503, buf504, buf507, buf508, buf511, buf512, buf515, buf516, buf519, buf520, buf523, buf524, buf527, buf528, buf531, buf532, buf535, buf536, buf539, buf540, buf543, buf544, buf547, buf548, buf551, buf552, buf555, buf556, buf559, buf560, buf563, buf564, buf567, buf568, buf571, buf572, buf575, buf576, buf579, buf580, buf583, buf584, buf587, buf588, buf591, buf592, buf594, buf595, buf599, buf600, buf603, buf604, buf607, buf608, buf611, buf612, buf615, buf616, buf619, buf620, buf624, primals_1, primals_2, primals_4, primals_5, primals_7, primals_8, primals_10, primals_11, primals_13, primals_14, primals_16, primals_17, primals_19, primals_20, primals_22, primals_23, primals_25, primals_26, primals_28, primals_29, primals_31, primals_32, primals_34, primals_35, primals_37, primals_38, primals_40, primals_41, primals_43, primals_44, primals_46, primals_47, primals_49, primals_50, primals_52, primals_53, primals_55, primals_56, primals_58, primals_59, primals_61, primals_62, primals_64, primals_65, primals_67, primals_68, primals_70, primals_71, primals_73, primals_74, primals_76, primals_77, primals_79, primals_80, primals_82, primals_83, primals_85, primals_86, primals_88, primals_89, primals_91, primals_92, primals_94, primals_95, primals_97, primals_98, primals_100, primals_101, primals_103, primals_104, primals_106, primals_107, primals_109, primals_110, primals_112, primals_113, primals_115, primals_116, primals_118, primals_119, primals_121, primals_122, primals_124, primals_125, primals_127, primals_128, primals_130, primals_131, primals_133, primals_134, primals_136, primals_137, primals_139, primals_140, primals_142, primals_143, primals_145, primals_146, primals_148, primals_149, primals_151, primals_152, primals_154, primals_155, primals_157, primals_158, primals_160, primals_161, primals_163, primals_164, primals_166, primals_167, primals_169, primals_170, primals_172, primals_173, primals_175, primals_176, primals_178, primals_179, primals_181, primals_182, primals_184, primals_185, primals_187, primals_188, primals_190, primals_191, primals_193, primals_194, primals_196, primals_197, primals_199, primals_200, primals_202, primals_203, primals_205, primals_206, primals_208, primals_209, primals_211, primals_212, primals_214, primals_215, primals_217, primals_218, primals_220, primals_221, primals_223, primals_224, primals_226, primals_227, primals_229, primals_230, primals_232, primals_233, primals_235, primals_236, primals_238, primals_239, primals_241, primals_242, primals_244, primals_245, primals_247, primals_248, primals_250, primals_251, primals_253, primals_254, primals_256, primals_257, primals_259, primals_260, primals_262, primals_263, primals_265, primals_266, primals_268, primals_269, primals_271, primals_272, primals_274, primals_275, primals_277, primals_278, primals_280, primals_281, primals_283, primals_284, primals_286, primals_287, primals_289, primals_290, primals_292, primals_293, primals_295, primals_296, primals_298, primals_299, primals_301, primals_302, primals_304, primals_305, primals_307, primals_308, primals_310, primals_311, primals_313, primals_314, primals_316, primals_317, primals_319, primals_320, primals_322, primals_323, primals_325, primals_326, primals_328, primals_329, primals_331, primals_332, primals_334, primals_335, primals_337, primals_338, primals_340, primals_341, primals_343, primals_344, primals_346, primals_347, primals_349, primals_350, primals_352, primals_353, primals_355, primals_356, primals_358, primals_359, primals_361, primals_362, primals_364, primals_365, primals_367, primals_368, primals_370, primals_371, primals_373, primals_374, primals_376, primals_377, primals_379, primals_380, primals_382, primals_383, primals_385, primals_386, primals_388, primals_389, primals_391, primals_392, primals_394, primals_395, primals_397, primals_398, primals_400, primals_401, primals_403, primals_404, primals_406, primals_407, primals_409, primals_410, primals_412, primals_413, primals_415, primals_416, primals_418, primals_419, primals_421, primals_422, primals_424, primals_425, primals_427, primals_428, primals_430, primals_431, primals_433, primals_434, primals_436, primals_437, primals_439, primals_440, primals_442, primals_443, primals_445, primals_446, primals_448, primals_449, primals_451, primals_452, primals_454, primals_455, primals_457, primals_458, primals_460, primals_461, primals_463, primals_464, primals_933, buf0, buf1, buf2, buf3, buf4, buf5, buf6, buf7, buf8, buf9, buf10, buf11, buf12, buf13, buf14, buf15, buf16, buf17, buf18, buf19, buf21, buf22, buf23, buf24, buf25, buf26, buf27, buf28, buf29, buf30, buf31, buf32, buf33, buf34, buf35, buf36, buf37, buf38, buf39, buf40, buf41, buf42, buf43, buf44, buf45, buf46, buf47, buf48, buf49, buf50, buf51, buf52, buf53, buf54, buf55, buf56, buf57, buf58, buf59, buf61, buf62, buf63, buf64, buf65, buf66, buf67, buf68, buf69, buf70, buf71, buf72, buf73, buf74, buf75, buf76, buf77, buf78, buf79, buf80, buf81, buf82, buf83, buf84, buf85, buf86, buf87, buf88, buf89, buf90, buf91, buf92, buf93, buf94, buf95, buf96, buf97, buf98, buf99, buf100, buf101, buf102, buf103, buf104, buf105, buf106, buf107, buf108, buf109, buf110, buf111, buf112, buf113, buf114, buf115, buf116, buf117, buf118, buf119, buf120, buf121, buf122, buf123, buf124, buf125, buf126, buf127, buf128, buf129, buf130, buf131, buf132, buf133, buf134, buf135, buf136, buf137, buf138, buf139, buf140, buf141, buf142, buf143, buf144, buf145, buf146, buf147, buf148, buf149, buf150, buf151, buf152, buf153, buf154, buf155, buf156, buf157, buf158, buf159, buf161, buf162, buf163, buf164, buf165, buf166, buf167, buf168, buf169, buf170, buf171, buf172, buf173, buf174, buf175, buf176, buf177, buf178, buf179, buf180, buf181, buf182, buf183, buf184, buf185, buf186, buf187, buf188, buf189, buf190, buf191, buf192, buf193, buf194, buf195, buf196, buf197, buf198, buf199, buf200, buf201, buf202, buf203, buf204, buf205, buf206, buf207, buf208, buf209, buf210, buf211, buf212, buf213, buf214, buf215, buf216, buf217, buf218, buf219, buf220, buf221, buf222, buf223, buf224, buf225, buf226, buf227, buf228, buf229, buf230, buf231, buf232, buf233, buf234, buf235, buf236, buf237, buf238, buf239, buf240, buf241, buf242, buf243, buf244, buf245, buf246, buf247, buf248, buf249, buf250, buf251, buf252, buf253, buf254, buf255, buf256, buf257, buf258, buf259, buf260, buf261, buf262, buf263, buf264, buf265, buf266, buf267, buf268, buf269, buf270, buf271, buf272, buf273, buf274, buf275, buf276, buf277, buf278, buf279, buf280, buf281, buf282, buf283, buf284, buf285, buf286, buf287, buf288, buf289, buf290, buf291, buf292, buf293, buf294, buf295, buf296, buf297, buf298, buf299, buf300, buf301, buf302, buf303, buf304, buf305, buf306, buf307, buf308, buf309, buf310, buf311, buf312, buf313, buf314, buf315, buf316, buf317, buf318, buf319, buf320, buf321, buf322, buf323, buf324, buf325, buf326, buf327, buf328, buf329, buf330, buf331, buf332, buf333, buf334, buf335, buf336, buf337, buf338, buf339, buf340, buf341, buf342, buf343, buf344, buf345, buf346, buf347, buf348, buf349, buf350, buf351, buf352, buf353, buf354, buf355, buf356, buf357, buf358, buf359, buf360, buf361, buf362, buf363, buf364, buf365, buf366, buf367, buf368, buf369, buf370, buf371, buf372, buf373, buf374, buf375, buf376, buf377, buf378, buf379, buf380, buf381, buf382, buf383, buf384, buf385, buf386, buf387, buf388, buf389, buf390, buf391, buf392, buf393, buf394, buf395, buf396, buf397, buf398, buf399, buf400, buf401, buf402, buf403, buf404, buf405, buf406, buf407, buf408, buf409, buf410, buf411, buf412, buf413, buf414, buf415, buf416, buf417, buf418, buf419, buf420, buf421, buf422, buf423, buf424, buf425, buf426, buf427, buf428, buf429, buf430, buf431, buf432, buf433, buf434, buf435, buf436, buf437, buf438, buf439, buf440, buf441, buf442, buf443, buf444, buf445, buf446, buf447, buf448, buf449, buf450, buf451, buf452, buf453, buf454, buf455, buf456, buf457, buf458, buf459, buf460, buf461, buf462, buf463, buf464, buf465, buf466, buf467, buf468, buf469, buf470, buf471, buf472, buf473, buf474, buf475, buf476, buf477, buf478, buf479, buf480, buf481, buf482, buf483, buf484, buf485, buf486, buf487, buf488, buf489, buf490, buf491, buf492, buf493, buf494, buf495, buf496, buf497, buf498, buf499, buf500, buf501, buf502, buf503, buf504, buf505, buf506, buf507, buf508, buf509, buf510, buf511, buf512, buf513, buf514, buf515, buf516, buf517, buf518, buf519, buf520, buf521, buf522, buf523, buf524, buf525, buf526, buf527, buf528, buf529, buf530, buf531, buf532, buf533, buf534, buf535, buf536, buf537, buf538, buf539, buf540, buf541, buf542, buf543, buf544, buf545, buf546, buf547, buf548, buf549, buf550, buf551, buf552, buf553, buf554, buf555, buf556, buf557, buf558, buf559, buf560, buf561, buf562, buf563, buf564, buf565, buf566, buf567, buf568, buf569, buf570, buf571, buf572, buf573, buf574, buf575, buf576, buf577, buf578, buf579, buf580, buf581, buf582, buf583, buf584, buf585, buf586, buf587, buf588, buf589, buf590, buf591, buf592, buf593, buf594, buf595, buf597, buf598, buf599, buf600, buf601, buf602, buf603, buf604, buf605, buf606, buf607, buf608, buf609, buf610, buf611, buf612, buf613, buf614, buf615, buf616, buf617, buf618, buf619, buf620, as_strided(buf623, (4, 2048), (2048, 1)), as_strided(primals_466, (1000, 2048), (2048, 1)), buf625, s0, )
NameError: name 's0' is not defined
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 515 ops
cuda train resnet18 ERROR:common:name 's0' is not defined
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/torchvision/torchvision/models/resnet.py", line 284, in forward
def forward(self, x: Tensor) -> Tensor:
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2340, in forward
return compiled_fn(full_args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 887, in g
return f(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1903, in debug_compiled_function
return compiled_function(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1718, in compiled_function
all_outs = CompiledFunction.apply(*args_with_synthetic_bases)
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 419, in apply
return super().apply(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1581, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 912, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 199, in run
return model(new_inputs)
File "/tmp/torchinductor_ezyang/h6/ch65dnxynpy6oom7cpioz6zksofxnbtklvtwq6zhytd2xx3ntzfm.py", line 971, in call
return (buf1, buf2, buf7, buf8, buf11, buf12, buf15, buf16, buf19, buf20, buf23, buf24, buf27, buf28, buf30, buf31, buf35, buf36, buf39, buf40, buf43, buf44, buf47, buf48, buf50, buf51, buf55, buf56, buf59, buf60, buf63, buf64, buf67, buf68, buf70, buf71, buf75, buf76, buf79, buf80, buf84, primals_1, primals_2, primals_4, primals_5, primals_7, primals_8, primals_10, primals_11, primals_13, primals_14, primals_16, primals_17, primals_19, primals_20, primals_22, primals_23, primals_25, primals_26, primals_28, primals_29, primals_31, primals_32, primals_34, primals_35, primals_37, primals_38, primals_40, primals_41, primals_43, primals_44, primals_46, primals_47, primals_49, primals_50, primals_52, primals_53, primals_55, primals_56, primals_58, primals_59, primals_123, buf0, buf1, buf2, buf3, buf4, buf5, buf6, buf7, buf8, buf9, buf10, buf11, buf12, buf13, buf14, buf15, buf16, buf17, buf18, buf19, buf20, buf21, buf22, buf23, buf24, buf25, buf26, buf27, buf28, buf29, buf30, buf31, buf33, buf34, buf35, buf36, buf37, buf38, buf39, buf40, buf41, buf42, buf43, buf44, buf45, buf46, buf47, buf48, buf49, buf50, buf51, buf53, buf54, buf55, buf56, buf57, buf58, buf59, buf60, buf61, buf62, buf63, buf64, buf65, buf66, buf67, buf68, buf69, buf70, buf71, buf73, buf74, buf75, buf76, buf77, buf78, buf79, buf80, as_strided(buf83, (4, 512), (512, 1)), as_strided(primals_61, (1000, 512), (512, 1)), buf85, s0, )
NameError: name 's0' is not defined
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 69 ops
cuda train resnet50 ERROR:common:name 's0' is not defined
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/torchvision/torchvision/models/resnet.py", line 284, in forward
def forward(self, x: Tensor) -> Tensor:
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2340, in forward
return compiled_fn(full_args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 887, in g
return f(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1903, in debug_compiled_function
return compiled_function(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1718, in compiled_function
all_outs = CompiledFunction.apply(*args_with_synthetic_bases)
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 419, in apply
return super().apply(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1581, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 912, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 199, in run
return model(new_inputs)
File "/tmp/torchinductor_ezyang/je/cje5fxxx2qqxbrfyj6vwpnkg25cht4gt5uzndo44m45r4vpioeb7.py", line 1569, in call
return (buf1, buf2, buf7, buf8, buf11, buf12, buf15, buf16, buf18, buf19, buf23, buf24, buf27, buf28, buf31, buf32, buf35, buf36, buf39, buf40, buf43, buf44, buf47, buf48, buf51, buf52, buf55, buf56, buf58, buf59, buf63, buf64, buf67, buf68, buf71, buf72, buf75, buf76, buf79, buf80, buf83, buf84, buf87, buf88, buf91, buf92, buf95, buf96, buf99, buf100, buf103, buf104, buf107, buf108, buf110, buf111, buf115, buf116, buf119, buf120, buf123, buf124, buf127, buf128, buf131, buf132, buf135, buf136, buf139, buf140, buf143, buf144, buf147, buf148, buf151, buf152, buf155, buf156, buf159, buf160, buf163, buf164, buf167, buf168, buf171, buf172, buf175, buf176, buf179, buf180, buf183, buf184, buf186, buf187, buf191, buf192, buf195, buf196, buf199, buf200, buf203, buf204, buf207, buf208, buf211, buf212, buf216, primals_1, primals_2, primals_4, primals_5, primals_7, primals_8, primals_10, primals_11, primals_13, primals_14, primals_16, primals_17, primals_19, primals_20, primals_22, primals_23, primals_25, primals_26, primals_28, primals_29, primals_31, primals_32, primals_34, primals_35, primals_37, primals_38, primals_40, primals_41, primals_43, primals_44, primals_46, primals_47, primals_49, primals_50, primals_52, primals_53, primals_55, primals_56, primals_58, primals_59, primals_61, primals_62, primals_64, primals_65, primals_67, primals_68, primals_70, primals_71, primals_73, primals_74, primals_76, primals_77, primals_79, primals_80, primals_82, primals_83, primals_85, primals_86, primals_88, primals_89, primals_91, primals_92, primals_94, primals_95, primals_97, primals_98, primals_100, primals_101, primals_103, primals_104, primals_106, primals_107, primals_109, primals_110, primals_112, primals_113, primals_115, primals_116, primals_118, primals_119, primals_121, primals_122, primals_124, primals_125, primals_127, primals_128, primals_130, primals_131, primals_133, primals_134, primals_136, primals_137, primals_139, primals_140, primals_142, primals_143, primals_145, primals_146, primals_148, primals_149, primals_151, primals_152, primals_154, primals_155, primals_157, primals_158, primals_321, buf0, buf1, buf2, buf3, buf4, buf5, buf6, buf7, buf8, buf9, buf10, buf11, buf12, buf13, buf14, buf15, buf16, buf17, buf18, buf19, buf21, buf22, buf23, buf24, buf25, buf26, buf27, buf28, buf29, buf30, buf31, buf32, buf33, buf34, buf35, buf36, buf37, buf38, buf39, buf40, buf41, buf42, buf43, buf44, buf45, buf46, buf47, buf48, buf49, buf50, buf51, buf52, buf53, buf54, buf55, buf56, buf57, buf58, buf59, buf61, buf62, buf63, buf64, buf65, buf66, buf67, buf68, buf69, buf70, buf71, buf72, buf73, buf74, buf75, buf76, buf77, buf78, buf79, buf80, buf81, buf82, buf83, buf84, buf85, buf86, buf87, buf88, buf89, buf90, buf91, buf92, buf93, buf94, buf95, buf96, buf97, buf98, buf99, buf100, buf101, buf102, buf103, buf104, buf105, buf106, buf107, buf108, buf109, buf110, buf111, buf113, buf114, buf115, buf116, buf117, buf118, buf119, buf120, buf121, buf122, buf123, buf124, buf125, buf126, buf127, buf128, buf129, buf130, buf131, buf132, buf133, buf134, buf135, buf136, buf137, buf138, buf139, buf140, buf141, buf142, buf143, buf144, buf145, buf146, buf147, buf148, buf149, buf150, buf151, buf152, buf153, buf154, buf155, buf156, buf157, buf158, buf159, buf160, buf161, buf162, buf163, buf164, buf165, buf166, buf167, buf168, buf169, buf170, buf171, buf172, buf173, buf174, buf175, buf176, buf177, buf178, buf179, buf180, buf181, buf182, buf183, buf184, buf185, buf186, buf187, buf189, buf190, buf191, buf192, buf193, buf194, buf195, buf196, buf197, buf198, buf199, buf200, buf201, buf202, buf203, buf204, buf205, buf206, buf207, buf208, buf209, buf210, buf211, buf212, as_strided(buf215, (4, 2048), (2048, 1)), as_strided(primals_160, (1000, 2048), (2048, 1)), buf217, s0, )
NameError: name 's0' is not defined
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 175 ops
cuda train resnet50_quantized_qat WARNING:common:fp64 golden ref were not generated for resnet50_quantized_qat. Setting accuracy check to cosine
[2022-12-12 06:55:23,622] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,633] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,658] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,684] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,692] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,713] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,721] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,740] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,746] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,766] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,771] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,789] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,794] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,801] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,820] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,827] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,847] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,853] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,871] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,877] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,883] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,902] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,909] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,928] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,934] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,952] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,958] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,965] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,986] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:23,994] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,017] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,025] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,046] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,052] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,072] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,077] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,085] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,109] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,116] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,135] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,141] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,160] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,165] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,172] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,191] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,198] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,217] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,224] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,242] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,248] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,255] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,275] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,283] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,303] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,309] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,329] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,335] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,342] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,363] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,379] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,402] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,802] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,824] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,830] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,850] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,856] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,865] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,886] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,894] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,915] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,922] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,943] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,950] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,957] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,979] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:24,986] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,008] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,016] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,036] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,043] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,051] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,082] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,089] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,112] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,120] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,140] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,147] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,154] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,176] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,184] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,205] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,213] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,234] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,240] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,247] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,270] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,278] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,300] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,309] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,331] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,339] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,347] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,383] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,395] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,420] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,430] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,454] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,462] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,486] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,493] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,505] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,529] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,537] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,561] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,569] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,594] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,603] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,614] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,638] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,646] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,671] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,678] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,719] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,728] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,735] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,740] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,744] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
[2022-12-12 06:55:25,750] torch._inductor.ir: [WARNING] Using FallbackKernel: aten._fused_moving_avg_obs_fq_helper_functional
ERROR:common:name 's0' is not defined
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/fx/graph_module.py", line 660, in call_wrapped
return self._wrapped_call(self, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/fx/graph_module.py", line 279, in __call__
raise e
File "/data/users/ezyang/a/pytorch/torch/fx/graph_module.py", line 269, in __call__
return super(self.cls, obj).__call__(*args, **kwargs) # type: ignore[misc]
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "<eval_with_key>.8", line 4, in forward
def forward(self, x : torch.Tensor) -> torch.Tensor:
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2340, in forward
return compiled_fn(full_args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 887, in g
return f(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1903, in debug_compiled_function
return compiled_function(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1718, in compiled_function
all_outs = CompiledFunction.apply(*args_with_synthetic_bases)
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 419, in apply
return super().apply(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1581, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 912, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 199, in run
return model(new_inputs)
File "/tmp/torchinductor_ezyang/ow/cowkjznhz3xsve5n5cxxr2yswzyg4lhbxds63lcrst2t7ykvrwxq.py", line 6536, in call
return (buf9, buf10, buf7, buf8, buf26, buf27, buf21, buf22, buf19, buf20, buf47, buf48, buf45, buf46, buf64, buf65, buf59, buf60, buf57, buf58, buf76, buf77, buf74, buf75, buf92, buf93, buf88, buf89, buf86, buf87, buf104, buf105, buf102, buf103, buf122, buf123, buf116, buf117, buf114, buf115, buf134, buf135, buf132, buf133, buf151, buf152, buf146, buf147, buf144, buf145, buf163, buf164, buf161, buf162, buf175, buf176, buf173, buf174, buf192, buf193, buf187, buf188, buf185, buf186, buf204, buf205, buf202, buf203, buf220, buf221, buf216, buf217, buf214, buf215, buf232, buf233, buf230, buf231, buf249, buf250, buf244, buf245, buf242, buf243, buf261, buf262, buf259, buf260, buf273, buf274, buf271, buf272, buf290, buf291, buf285, buf286, buf283, buf284, buf302, buf303, buf300, buf301, buf318, buf319, buf314, buf315, buf312, buf313, buf330, buf331, buf328, buf329, buf347, buf348, buf342, buf343, buf340, buf341, buf359, buf360, buf357, buf358, buf371, buf372, buf369, buf370, buf389, buf390, buf383, buf384, buf381, buf382, buf401, buf402, buf399, buf400, buf417, buf418, buf413, buf414, buf411, buf412, buf429, buf430, buf427, buf428, buf447, buf448, buf441, buf442, buf439, buf440, buf459, buf460, buf457, buf458, buf476, buf477, buf471, buf472, buf469, buf470, buf488, buf489, buf486, buf487, buf500, buf501, buf498, buf499, buf517, buf518, buf512, buf513, buf510, buf511, buf529, buf530, buf527, buf528, buf545, buf546, buf541, buf542, buf539, buf540, buf557, buf558, buf555, buf556, buf574, buf575, buf569, buf570, buf567, buf568, buf586, buf587, buf584, buf585, buf598, buf599, buf596, buf597, buf615, buf616, buf610, buf611, buf608, buf609, buf627, buf628, buf625, buf626, buf643, buf644, buf639, buf640, buf637, buf638, buf655, buf656, buf653, buf654, buf672, buf673, buf667, buf668, buf665, buf666, buf684, buf685, buf682, buf683, buf696, buf697, buf694, buf695, buf713, buf714, buf708, buf709, buf706, buf707, buf725, buf726, buf723, buf724, buf741, buf742, buf737, buf738, buf735, buf736, buf753, buf754, buf751, buf752, buf770, buf771, buf765, buf766, buf763, buf764, buf782, buf783, buf780, buf781, buf794, buf795, buf792, buf793, buf811, buf812, buf806, buf807, buf804, buf805, buf823, buf824, buf821, buf822, buf839, buf840, buf835, buf836, buf833, buf834, buf851, buf852, buf849, buf850, buf869, buf870, buf863, buf864, buf861, buf862, buf881, buf882, buf879, buf880, buf898, buf899, buf893, buf894, buf891, buf892, buf910, buf911, buf908, buf909, buf922, buf923, buf920, buf921, buf939, buf940, buf934, buf935, buf932, buf933, buf951, buf952, buf949, buf950, buf967, buf968, buf963, buf964, buf961, buf962, buf979, buf980, buf977, buf978, buf996, buf997, buf991, buf992, buf989, buf990, buf1008, buf1009, buf1006, buf1007, buf1020, buf1021, buf1018, buf1019, buf1037, buf1038, buf1032, buf1033, buf1030, buf1031, buf1049, buf1050, buf1047, buf1048, buf1065, buf1066, buf1061, buf1062, buf1059, buf1060, buf1077, buf1078, buf1075, buf1076, buf1094, buf1095, buf1089, buf1090, buf1087, buf1088, buf1106, buf1107, buf1104, buf1105, buf1118, buf1119, buf1116, buf1117, buf1135, buf1136, buf1130, buf1131, buf1128, buf1129, buf1147, buf1148, buf1145, buf1146, buf1163, buf1164, buf1159, buf1160, buf1157, buf1158, buf1175, buf1176, buf1173, buf1174, buf1192, buf1193, buf1187, buf1188, buf1185, buf1186, buf1204, buf1205, buf1202, buf1203, buf1216, buf1217, buf1214, buf1215, buf1233, buf1234, buf1228, buf1229, buf1226, buf1227, buf1245, buf1246, buf1243, buf1244, buf1261, buf1262, buf1257, buf1258, buf1255, buf1256, buf1273, buf1274, buf1271, buf1272, buf1290, buf1291, buf1285, buf1286, buf1283, buf1284, buf1302, buf1303, buf1300, buf1301, buf1314, buf1315, buf1312, buf1313, buf1331, buf1332, buf1326, buf1327, buf1324, buf1325, buf1343, buf1344, buf1341, buf1342, buf1359, buf1360, buf1355, buf1356, buf1353, buf1354, buf1371, buf1372, buf1369, buf1370, buf1388, buf1389, buf1383, buf1384, buf1381, buf1382, buf1400, buf1401, buf1398, buf1399, buf1412, buf1413, buf1410, buf1411, buf1429, buf1430, buf1424, buf1425, buf1422, buf1423, buf1441, buf1442, buf1439, buf1440, buf1457, buf1458, buf1453, buf1454, buf1451, buf1452, buf1469, buf1470, buf1467, buf1468, buf1487, buf1488, buf1481, buf1482, buf1479, buf1480, buf1499, buf1500, buf1497, buf1498, buf1516, buf1517, buf1511, buf1512, buf1509, buf1510, buf1528, buf1529, buf1526, buf1527, buf1540, buf1541, buf1538, buf1539, buf1557, buf1558, buf1552, buf1553, buf1550, buf1551, buf1569, buf1570, buf1567, buf1568, buf1585, buf1586, buf1581, buf1582, buf1579, buf1580, buf1597, buf1598, buf1595, buf1596, buf1614, buf1615, buf1609, buf1610, buf1607, buf1608, buf1626, buf1627, buf1624, buf1625, buf1638, buf1639, buf1636, buf1637, buf1655, buf1656, buf1650, buf1651, buf1648, buf1649, buf1667, buf1668, buf1665, buf1666, buf1683, buf1684, buf1679, buf1680, buf1677, buf1678, buf1695, buf1696, buf1693, buf1694, buf1712, buf1713, buf1707, buf1708, buf1705, buf1706, buf1724, buf1725, buf1722, buf1723, buf1752, buf1753, buf1750, buf1751, buf1763, buf1764, buf1761, buf1762, buf1775, buf1776, buf1773, buf1774, buf1771, primals_1, primals_2, primals_4, primals_5, primals_7, primals_8, primals_10, primals_11, primals_13, primals_14, primals_16, primals_17, primals_19, primals_20, primals_22, primals_23, primals_25, primals_26, primals_28, primals_29, primals_31, primals_32, primals_34, primals_35, primals_37, primals_38, primals_40, primals_41, primals_43, primals_44, primals_46, primals_47, primals_49, primals_50, primals_52, primals_53, primals_55, primals_56, primals_58, primals_59, primals_61, primals_62, primals_64, primals_65, primals_67, primals_68, primals_70, primals_71, primals_73, primals_74, primals_76, primals_77, primals_79, primals_80, primals_82, primals_83, primals_85, primals_86, primals_88, primals_89, primals_91, primals_92, primals_94, primals_95, primals_97, primals_98, primals_100, primals_101, primals_103, primals_104, primals_106, primals_107, primals_109, primals_110, primals_112, primals_113, primals_115, primals_116, primals_118, primals_119, primals_121, primals_122, primals_124, primals_125, primals_127, primals_128, primals_130, primals_131, primals_133, primals_134, primals_136, primals_137, primals_139, primals_140, primals_142, primals_143, primals_145, primals_146, primals_148, primals_149, primals_151, primals_152, primals_154, primals_155, primals_157, primals_158, primals_170, primals_187, primals_204, primals_221, primals_238, primals_262, primals_279, primals_296, primals_320, primals_337, primals_354, primals_378, primals_395, primals_412, primals_429, primals_453, primals_470, primals_487, primals_511, primals_528, primals_545, primals_569, primals_586, primals_603, primals_627, primals_644, primals_661, primals_678, primals_702, primals_719, primals_736, primals_760, primals_777, primals_794, primals_818, primals_835, primals_852, primals_876, primals_893, primals_910, primals_934, primals_951, primals_968, primals_992, primals_1009, primals_1026, primals_1043, primals_1067, primals_1084, primals_1101, primals_1125, primals_1142, primals_1159, buf5, buf17, buf18, buf25, buf26, buf34, buf35, buf41, buf43, buf44, buf61, buf56, buf63, buf64, buf72, buf73, buf84, buf85, buf91, buf92, buf100, buf101, buf118, buf113, buf121, buf122, buf131, buf148, buf143, buf150, buf151, buf160, buf171, buf172, buf189, buf184, buf191, buf192, buf200, buf201, buf212, buf213, buf219, buf220, buf228, buf229, buf246, buf241, buf248, buf249, buf258, buf269, buf270, buf287, buf282, buf289, buf290, buf298, buf299, buf310, buf311, buf317, buf318, buf326, buf327, buf344, buf339, buf346, buf347, buf356, buf367, buf368, buf385, buf380, buf388, buf389, buf397, buf398, buf409, buf410, buf416, buf417, buf425, buf426, buf443, buf438, buf446, buf447, buf456, buf473, buf468, buf475, buf476, buf485, buf496, buf497, buf514, buf509, buf516, buf517, buf525, buf526, buf537, buf538, buf544, buf545, buf553, buf554, buf571, buf566, buf573, buf574, buf583, buf594, buf595, buf612, buf607, buf614, buf615, buf623, buf624, buf635, buf636, buf642, buf643, buf651, buf652, buf669, buf664, buf671, buf672, buf681, buf692, buf693, buf710, buf705, buf712, buf713, buf721, buf722, buf733, buf734, buf740, buf741, buf749, buf750, buf767, buf762, buf769, buf770, buf779, buf790, buf791, buf808, buf803, buf810, buf811, buf819, buf820, buf831, buf832, buf838, buf839, buf847, buf848, buf865, buf860, buf868, buf869, buf878, buf895, buf890, buf897, buf898, buf907, buf918, buf919, buf936, buf931, buf938, buf939, buf947, buf948, buf959, buf960, buf966, buf967, buf975, buf976, buf993, buf988, buf995, buf996, buf1005, buf1016, buf1017, buf1034, buf1029, buf1036, buf1037, buf1045, buf1046, buf1057, buf1058, buf1064, buf1065, buf1073, buf1074, buf1091, buf1086, buf1093, buf1094, buf1103, buf1114, buf1115, buf1132, buf1127, buf1134, buf1135, buf1143, buf1144, buf1155, buf1156, buf1162, buf1163, buf1171, buf1172, buf1189, buf1184, buf1191, buf1192, buf1201, buf1212, buf1213, buf1230, buf1225, buf1232, buf1233, buf1241, buf1242, buf1253, buf1254, buf1260, buf1261, buf1269, buf1270, buf1287, buf1282, buf1289, buf1290, buf1299, buf1310, buf1311, buf1328, buf1323, buf1330, buf1331, buf1339, buf1340, buf1351, buf1352, buf1358, buf1359, buf1367, buf1368, buf1385, buf1380, buf1387, buf1388, buf1397, buf1408, buf1409, buf1426, buf1421, buf1428, buf1429, buf1437, buf1438, buf1449, buf1450, buf1456, buf1457, buf1465, buf1466, buf1483, buf1478, buf1486, buf1487, buf1496, buf1513, buf1508, buf1515, buf1516, buf1525, buf1536, buf1537, buf1554, buf1549, buf1556, buf1557, buf1565, buf1566, buf1577, buf1578, buf1584, buf1585, buf1593, buf1594, buf1611, buf1606, buf1613, buf1614, buf1623, buf1634, buf1635, buf1652, buf1647, buf1654, buf1655, buf1663, buf1664, buf1675, buf1676, buf1682, buf1683, buf1691, buf1692, buf1709, buf1704, buf1711, buf1712, buf1721, buf1733, buf1742, buf1748, buf1749, buf1760, buf1772, as_strided(buf1759, (1000, 2048), (2048, 1)), buf1777, buf1778, buf1779, buf1780, buf1781, buf1782, buf1783, buf1784, buf1785, buf1786, buf1787, buf1788, buf1789, buf1790, buf1791, buf1792, buf1793, buf1794, buf1795, buf1796, buf1797, buf1798, buf1799, buf1800, buf1801, buf1802, buf1803, buf1804, buf1805, buf1806, buf1807, buf1808, buf1809, buf1810, buf1811, buf1812, buf1813, buf1814, buf1815, buf1816, buf1817, buf1818, buf1819, buf1820, buf1821, buf1822, buf1823, buf1824, buf1825, s0, )
NameError: name 's0' is not defined
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 163 ops
cuda train resnext50_32x4d ERROR:common:name 's0' is not defined
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/torchvision/torchvision/models/resnet.py", line 284, in forward
def forward(self, x: Tensor) -> Tensor:
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2340, in forward
return compiled_fn(full_args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 887, in g
return f(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1903, in debug_compiled_function
return compiled_function(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1718, in compiled_function
all_outs = CompiledFunction.apply(*args_with_synthetic_bases)
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 419, in apply
return super().apply(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1581, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 912, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 199, in run
return model(new_inputs)
File "/tmp/torchinductor_ezyang/zx/czxdjqzfafoubcx2lj4sneafh74fnf7zsvzshykslm335hitakww.py", line 1569, in call
return (buf1, buf2, buf7, buf8, buf11, buf12, buf15, buf16, buf18, buf19, buf23, buf24, buf27, buf28, buf31, buf32, buf35, buf36, buf39, buf40, buf43, buf44, buf47, buf48, buf51, buf52, buf55, buf56, buf58, buf59, buf63, buf64, buf67, buf68, buf71, buf72, buf75, buf76, buf79, buf80, buf83, buf84, buf87, buf88, buf91, buf92, buf95, buf96, buf99, buf100, buf103, buf104, buf107, buf108, buf110, buf111, buf115, buf116, buf119, buf120, buf123, buf124, buf127, buf128, buf131, buf132, buf135, buf136, buf139, buf140, buf143, buf144, buf147, buf148, buf151, buf152, buf155, buf156, buf159, buf160, buf163, buf164, buf167, buf168, buf171, buf172, buf175, buf176, buf179, buf180, buf183, buf184, buf186, buf187, buf191, buf192, buf195, buf196, buf199, buf200, buf203, buf204, buf207, buf208, buf211, buf212, buf216, primals_1, primals_2, primals_4, primals_5, primals_7, primals_8, primals_10, primals_11, primals_13, primals_14, primals_16, primals_17, primals_19, primals_20, primals_22, primals_23, primals_25, primals_26, primals_28, primals_29, primals_31, primals_32, primals_34, primals_35, primals_37, primals_38, primals_40, primals_41, primals_43, primals_44, primals_46, primals_47, primals_49, primals_50, primals_52, primals_53, primals_55, primals_56, primals_58, primals_59, primals_61, primals_62, primals_64, primals_65, primals_67, primals_68, primals_70, primals_71, primals_73, primals_74, primals_76, primals_77, primals_79, primals_80, primals_82, primals_83, primals_85, primals_86, primals_88, primals_89, primals_91, primals_92, primals_94, primals_95, primals_97, primals_98, primals_100, primals_101, primals_103, primals_104, primals_106, primals_107, primals_109, primals_110, primals_112, primals_113, primals_115, primals_116, primals_118, primals_119, primals_121, primals_122, primals_124, primals_125, primals_127, primals_128, primals_130, primals_131, primals_133, primals_134, primals_136, primals_137, primals_139, primals_140, primals_142, primals_143, primals_145, primals_146, primals_148, primals_149, primals_151, primals_152, primals_154, primals_155, primals_157, primals_158, primals_321, buf0, buf1, buf2, buf3, buf4, buf5, buf6, buf7, buf8, buf9, buf10, buf11, buf12, buf13, buf14, buf15, buf16, buf17, buf18, buf19, buf21, buf22, buf23, buf24, buf25, buf26, buf27, buf28, buf29, buf30, buf31, buf32, buf33, buf34, buf35, buf36, buf37, buf38, buf39, buf40, buf41, buf42, buf43, buf44, buf45, buf46, buf47, buf48, buf49, buf50, buf51, buf52, buf53, buf54, buf55, buf56, buf57, buf58, buf59, buf61, buf62, buf63, buf64, buf65, buf66, buf67, buf68, buf69, buf70, buf71, buf72, buf73, buf74, buf75, buf76, buf77, buf78, buf79, buf80, buf81, buf82, buf83, buf84, buf85, buf86, buf87, buf88, buf89, buf90, buf91, buf92, buf93, buf94, buf95, buf96, buf97, buf98, buf99, buf100, buf101, buf102, buf103, buf104, buf105, buf106, buf107, buf108, buf109, buf110, buf111, buf113, buf114, buf115, buf116, buf117, buf118, buf119, buf120, buf121, buf122, buf123, buf124, buf125, buf126, buf127, buf128, buf129, buf130, buf131, buf132, buf133, buf134, buf135, buf136, buf137, buf138, buf139, buf140, buf141, buf142, buf143, buf144, buf145, buf146, buf147, buf148, buf149, buf150, buf151, buf152, buf153, buf154, buf155, buf156, buf157, buf158, buf159, buf160, buf161, buf162, buf163, buf164, buf165, buf166, buf167, buf168, buf169, buf170, buf171, buf172, buf173, buf174, buf175, buf176, buf177, buf178, buf179, buf180, buf181, buf182, buf183, buf184, buf185, buf186, buf187, buf189, buf190, buf191, buf192, buf193, buf194, buf195, buf196, buf197, buf198, buf199, buf200, buf201, buf202, buf203, buf204, buf205, buf206, buf207, buf208, buf209, buf210, buf211, buf212, as_strided(buf215, (4, 2048), (2048, 1)), as_strided(primals_160, (1000, 2048), (2048, 1)), buf217, s0, )
NameError: name 's0' is not defined
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 175 ops
cuda train shufflenet_v2_x1_0 ERROR:common:name 's0' is not defined
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/torchvision/torchvision/models/shufflenetv2.py", line 165, in forward
def forward(self, x: Tensor) -> Tensor:
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2340, in forward
return compiled_fn(full_args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 887, in g
return f(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1903, in debug_compiled_function
return compiled_function(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1718, in compiled_function
all_outs = CompiledFunction.apply(*args_with_synthetic_bases)
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 419, in apply
return super().apply(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1581, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 912, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 199, in run
return model(new_inputs)
File "/tmp/torchinductor_ezyang/wa/cwachfmlmnf5dzhuk226njm66oymuij2c4yd5cfwogyjcqeb2zlz.py", line 1833, in call
return (buf1, buf2, buf7, buf8, buf11, buf12, buf15, buf16, buf19, buf20, buf23, buf24, buf29, buf30, buf33, buf34, buf37, buf38, buf44, buf45, buf48, buf49, buf52, buf53, buf59, buf60, buf63, buf64, buf67, buf68, buf74, buf75, buf78, buf79, buf82, buf83, buf86, buf87, buf90, buf91, buf96, buf97, buf100, buf101, buf104, buf105, buf111, buf112, buf115, buf116, buf119, buf120, buf126, buf127, buf130, buf131, buf134, buf135, buf141, buf142, buf145, buf146, buf149, buf150, buf156, buf157, buf160, buf161, buf164, buf165, buf171, buf172, buf175, buf176, buf179, buf180, buf186, buf187, buf190, buf191, buf194, buf195, buf201, buf202, buf205, buf206, buf209, buf210, buf213, buf214, buf217, buf218, buf223, buf224, buf227, buf228, buf231, buf232, buf238, buf239, buf242, buf243, buf246, buf247, buf253, buf254, buf257, buf258, buf261, buf262, buf268, buf269, buf273, primals_1, primals_2, primals_4, primals_5, primals_7, primals_8, primals_10, primals_11, primals_13, primals_14, primals_16, primals_17, primals_19, primals_20, primals_22, primals_23, primals_25, primals_26, primals_28, primals_29, primals_31, primals_32, primals_34, primals_35, primals_37, primals_38, primals_40, primals_41, primals_43, primals_44, primals_46, primals_47, primals_49, primals_50, primals_52, primals_53, primals_55, primals_56, primals_58, primals_59, primals_61, primals_62, primals_64, primals_65, primals_67, primals_68, primals_70, primals_71, primals_73, primals_74, primals_76, primals_77, primals_79, primals_80, primals_82, primals_83, primals_85, primals_86, primals_88, primals_89, primals_91, primals_92, primals_94, primals_95, primals_97, primals_98, primals_100, primals_101, primals_103, primals_104, primals_106, primals_107, primals_109, primals_110, primals_112, primals_113, primals_115, primals_116, primals_118, primals_119, primals_121, primals_122, primals_124, primals_125, primals_127, primals_128, primals_130, primals_131, primals_133, primals_134, primals_136, primals_137, primals_139, primals_140, primals_142, primals_143, primals_145, primals_146, primals_148, primals_149, primals_151, primals_152, primals_154, primals_155, primals_157, primals_158, primals_160, primals_161, primals_163, primals_164, primals_166, primals_167, primals_339, buf0, buf1, buf2, buf3, buf4, buf5, buf6, buf7, buf8, buf9, buf10, buf11, buf12, buf14, buf15, buf16, buf17, buf18, buf19, buf20, buf21, buf22, buf23, buf24, buf27, buf28, buf29, buf30, buf31, buf32, buf33, buf34, buf35, buf36, buf37, buf38, buf42, buf43, buf44, buf45, buf46, buf47, buf48, buf49, buf50, buf51, buf52, buf53, buf57, buf58, buf59, buf60, buf61, buf62, buf63, buf64, buf65, buf66, buf67, buf68, buf72, buf73, buf74, buf75, buf76, buf77, buf78, buf79, buf81, buf82, buf83, buf84, buf85, buf86, buf87, buf88, buf89, buf90, buf91, buf94, buf95, buf96, buf97, buf98, buf99, buf100, buf101, buf102, buf103, buf104, buf105, buf109, buf110, buf111, buf112, buf113, buf114, buf115, buf116, buf117, buf118, buf119, buf120, buf124, buf125, buf126, buf127, buf128, buf129, buf130, buf131, buf132, buf133, buf134, buf135, buf139, buf140, buf141, buf142, buf143, buf144, buf145, buf146, buf147, buf148, buf149, buf150, buf154, buf155, buf156, buf157, buf158, buf159, buf160, buf161, buf162, buf163, buf164, buf165, buf169, buf170, buf171, buf172, buf173, buf174, buf175, buf176, buf177, buf178, buf179, buf180, buf184, buf185, buf186, buf187, buf188, buf189, buf190, buf191, buf192, buf193, buf194, buf195, buf199, buf200, buf201, buf202, buf203, buf204, buf205, buf206, buf208, buf209, buf210, buf211, buf212, buf213, buf214, buf215, buf216, buf217, buf218, buf221, buf222, buf223, buf224, buf225, buf226, buf227, buf228, buf229, buf230, buf231, buf232, buf236, buf237, buf238, buf239, buf240, buf241, buf242, buf243, buf244, buf245, buf246, buf247, buf251, buf252, buf253, buf254, buf255, buf256, buf257, buf258, buf259, buf260, buf261, buf262, buf266, buf267, buf268, buf269, buf272, as_strided(primals_169, (1000, 1024), (1024, 1)), buf274, buf275, buf276, buf277, buf278, buf279, buf280, buf281, buf282, buf283, buf284, buf285, buf286, buf287, buf288, buf289, buf290, buf291, buf292, buf293, s0, 28, 28, 14, 14, 7, 7, )
NameError: name 's0' is not defined
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 367 ops
/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/gym/core.py:317: DeprecationWarning: WARN: Initializing wrapper in old step API which returns one bool instead of two. It is recommended to set `new_step_api=True` to use new step API. This will be the default behaviour in future.
deprecation(
/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/gym/wrappers/step_api_compatibility.py:39: DeprecationWarning: WARN: Initializing environment in old step API which returns one bool instead of two. It is recommended to set `new_step_api=True` to use new step API. This will be the default behaviour in future.
deprecation(
cuda train soft_actor_critic ERROR:common:compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 440, in codegen
self.scheduler.codegen()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/scheduler.py", line 1129, in codegen
self.get_backend(device).codegen_nodes(node.get_nodes())
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1284, in codegen_nodes
return self.codegen_node_schedule(node_schedule, numel, rnumel)
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1336, in codegen_node_schedule
src_code = kernel.codegen_kernel()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1035, in codegen_kernel
"signature": dict(enumerate(map(signature_of, signature))),
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 57, in signature_of
return JITFunction._key_of(V.graph.sizevars.size_hint(arg.expr))
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 367, in size_hint
return int(out)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 360, in <graph break in forward_and_backward_pass>
loss = self.compute_loss(pred)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 350, in compute_loss
return reduce_to_scalar_loss(pred)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/testing.py", line 105, in reduce_to_scalar_loss
return out.mean.sum()
File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/models/soft_actor_critic/nets.py", line 255, in mean
mu = tr(mu)
File "/data/users/ezyang/a/pytorch/torch/distributions/transforms.py", line 156, in __call__
y = self._call(x)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 476, in compile_subgraph
self.compile_and_call_fx_graph(tx, list(reversed(stack_values)), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 2 graph(s) covering 20 ops
cuda train speech_transformer ERROR:common:compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 437, in codegen
self.init_wrapper_code()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 431, in init_wrapper_code
self.wrapper_code = WrapperCodeGen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 302, in __init__
self.write_prefix()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 335, in write_prefix
V.graph.sizevars.codegen(self.wrapper_call, V.graph.graph_inputs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 494, in codegen
assert not needed
AssertionError
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/models/speech_transformer/speech_transformer/transformer/transformer.py", line 28, in forward
encoder_padded_outputs, *_ = self.encoder(padded_input, input_lengths)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/models/speech_transformer/speech_transformer/transformer/encoder.py", line 48, in forward
non_pad_mask = get_non_pad_mask(padded_input, input_lengths=input_lengths)
File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/models/speech_transformer/speech_transformer/transformer/encoder.py", line 50, in <graph break in forward>
slf_attn_mask = get_attn_pad_mask(padded_input, input_lengths, length)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 332, in wrapper
self.output.compile_subgraph(self, reason=reason)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 3 graph(s) covering 11 ops
cuda train squeezenet1_1 ERROR:common:name 's0' is not defined
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/torchvision/torchvision/models/squeezenet.py", line 94, in forward
def forward(self, x: torch.Tensor) -> torch.Tensor:
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2340, in forward
return compiled_fn(full_args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 887, in g
return f(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1903, in debug_compiled_function
return compiled_function(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1718, in compiled_function
all_outs = CompiledFunction.apply(*args_with_synthetic_bases)
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 419, in apply
return super().apply(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1581, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 912, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 199, in run
return model(new_inputs)
File "/tmp/torchinductor_ezyang/xk/cxkeuuj4s5eu77f77phnxynep4k5uigmbjpwkjhppcyvtzsxjzmi.py", line 678, in call
return (as_strided(buf66, (4, 1000), (1000, 1)), primals_1, primals_3, primals_5, primals_7, primals_9, primals_11, primals_13, primals_15, primals_17, primals_19, primals_21, primals_23, primals_25, primals_27, primals_29, primals_31, primals_33, primals_35, primals_37, primals_39, primals_41, primals_43, primals_45, primals_47, primals_49, primals_51, primals_53, buf1, buf2, buf3, buf5, buf10, buf12, buf17, buf18, buf19, buf21, buf26, buf28, buf33, buf34, buf35, buf37, buf42, buf44, buf49, buf51, buf56, buf58, buf63, buf67, buf68, buf69, buf70, buf71, buf72, buf73, buf74, buf75, buf76, buf77, buf78, buf79, buf80, buf81, buf82, buf83, s0, 13, 13, )
NameError: name 's0' is not defined
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 66 ops
cuda train tacotron2 ERROR:common:compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 437, in codegen
self.init_wrapper_code()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 431, in init_wrapper_code
self.wrapper_code = WrapperCodeGen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 302, in __init__
self.write_prefix()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 335, in write_prefix
V.graph.sizevars.codegen(self.wrapper_call, V.graph.graph_inputs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 494, in codegen
assert not needed
AssertionError
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 332, in wrapper
self.output.compile_subgraph(self, reason=reason)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 4 ops
cuda train timm_efficientdet ERROR:common:
from user code:
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/effdet/efficientdet.py", line 211, in forward
input_node = resample(input_node)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/effdet/efficientdet.py", line 134, in forward
return F.interpolate(
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 1087, in run_node
return node.target(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/nn/functional.py", line 3924, in interpolate
return torch._C._nn.upsample_nearest2d(input, output_size, scale_factors)
RuntimeError: Cannot call sizes() on tensor with symbolic sizes/strides
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 1046, in get_fake_value
return wrap_fake_exception(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 712, in wrap_fake_exception
return fn()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 1047, in <lambda>
lambda: run_node(tx.output, node, args, kwargs, nnmodule)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 1096, in run_node
raise RuntimeError(
RuntimeError: Failed running call_function <function interpolate at 0x7f1cbc6a6ee0>(*(FakeTensor(FakeTensor(..., device='meta',
size=(s0, 88, ceiling(ceiling(ceiling(ceiling(ceiling(ceiling(ceiling(s2/2)/2)/2)/2)/2)/2)/2), ceiling(ceiling(ceiling(ceiling(ceiling(ceiling(ceiling(s2/2)/2)/2)/2)/2)/2)/2)),
grad_fn=<MaxPool2DWithIndicesBackward0>), cuda:0), (10, 10), None, 'nearest', None), **{'recompute_scale_factor': False}):
Cannot call sizes() on tensor with symbolic sizes/strides
(scroll up for backtrace)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 305, in wrapper
return inner_fn(self, inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 956, in CALL_FUNCTION
self.call_function(fn, args, {})
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 430, in call_function
self.push(fn.call_function(self, args, kwargs))
File "/data/users/ezyang/a/pytorch/torch/_dynamo/variables/nn_module.py", line 220, in call_function
return tx.inline_user_function_return(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 466, in inline_user_function_return
result = InliningInstructionTranslator.inline_call(self, fn, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1750, in inline_call
return cls.inline_call_(parent, func, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1805, in inline_call_
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 305, in wrapper
return inner_fn(self, inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 956, in CALL_FUNCTION
self.call_function(fn, args, {})
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 430, in call_function
self.push(fn.call_function(self, args, kwargs))
File "/data/users/ezyang/a/pytorch/torch/_dynamo/variables/nn_module.py", line 220, in call_function
return tx.inline_user_function_return(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 466, in inline_user_function_return
result = InliningInstructionTranslator.inline_call(self, fn, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1750, in inline_call
return cls.inline_call_(parent, func, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1805, in inline_call_
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 305, in wrapper
return inner_fn(self, inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 956, in CALL_FUNCTION
self.call_function(fn, args, {})
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 430, in call_function
self.push(fn.call_function(self, args, kwargs))
File "/data/users/ezyang/a/pytorch/torch/_dynamo/variables/nn_module.py", line 220, in call_function
return tx.inline_user_function_return(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 466, in inline_user_function_return
result = InliningInstructionTranslator.inline_call(self, fn, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1750, in inline_call
return cls.inline_call_(parent, func, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1805, in inline_call_
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 305, in wrapper
return inner_fn(self, inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 956, in CALL_FUNCTION
self.call_function(fn, args, {})
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 430, in call_function
self.push(fn.call_function(self, args, kwargs))
File "/data/users/ezyang/a/pytorch/torch/_dynamo/variables/nn_module.py", line 220, in call_function
return tx.inline_user_function_return(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 466, in inline_user_function_return
result = InliningInstructionTranslator.inline_call(self, fn, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1750, in inline_call
return cls.inline_call_(parent, func, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1805, in inline_call_
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 305, in wrapper
return inner_fn(self, inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 956, in CALL_FUNCTION
self.call_function(fn, args, {})
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 430, in call_function
self.push(fn.call_function(self, args, kwargs))
File "/data/users/ezyang/a/pytorch/torch/_dynamo/variables/nn_module.py", line 220, in call_function
return tx.inline_user_function_return(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 466, in inline_user_function_return
result = InliningInstructionTranslator.inline_call(self, fn, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1750, in inline_call
return cls.inline_call_(parent, func, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1805, in inline_call_
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 305, in wrapper
return inner_fn(self, inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 956, in CALL_FUNCTION
self.call_function(fn, args, {})
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 430, in call_function
self.push(fn.call_function(self, args, kwargs))
File "/data/users/ezyang/a/pytorch/torch/_dynamo/variables/nn_module.py", line 220, in call_function
return tx.inline_user_function_return(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 466, in inline_user_function_return
result = InliningInstructionTranslator.inline_call(self, fn, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1750, in inline_call
return cls.inline_call_(parent, func, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1805, in inline_call_
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 305, in wrapper
return inner_fn(self, inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 956, in CALL_FUNCTION
self.call_function(fn, args, {})
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 430, in call_function
self.push(fn.call_function(self, args, kwargs))
File "/data/users/ezyang/a/pytorch/torch/_dynamo/variables/nn_module.py", line 182, in call_function
tx.call_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 430, in call_function
self.push(fn.call_function(self, args, kwargs))
File "/data/users/ezyang/a/pytorch/torch/_dynamo/variables/nn_module.py", line 220, in call_function
return tx.inline_user_function_return(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 466, in inline_user_function_return
result = InliningInstructionTranslator.inline_call(self, fn, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1750, in inline_call
return cls.inline_call_(parent, func, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1805, in inline_call_
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 305, in wrapper
return inner_fn(self, inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1005, in CALL_FUNCTION_KW
self.call_function(fn, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 430, in call_function
self.push(fn.call_function(self, args, kwargs))
File "/data/users/ezyang/a/pytorch/torch/_dynamo/variables/torch.py", line 444, in call_function
tensor_variable = wrap_fx_proxy(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/variables/builder.py", line 729, in wrap_fx_proxy
return wrap_fx_proxy_cls(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/variables/builder.py", line 766, in wrap_fx_proxy_cls
example_value = get_fake_value(proxy.node, tx)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 1066, in get_fake_value
raise TorchRuntimeError() from e
torch._dynamo.exc.TorchRuntimeError:
from user code:
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/effdet/efficientdet.py", line 211, in forward
input_node = resample(input_node)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/effdet/efficientdet.py", line 134, in forward
return F.interpolate(
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 0 ops
cuda train timm_efficientnet ERROR:common:compile_fx raised RuntimeError: Trying to extract a concrete int out of a symbolic int
While executing %primals_214 : [#users=1] = placeholder[target=primals_214]
Original traceback:
None
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/cache.py", line 70, in wrapper
retval = cfunc(*args, **kwargs)
TypeError: unhashable type: 'SymInt'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 231, in placeholder
sizes, strides = self.static_sizes_strides(example)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 62, in static_sizes_strides
size = [sympy.Integer(i) for i in ex.size()]
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 62, in <listcomp>
size = [sympy.Integer(i) for i in ex.size()]
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/cache.py", line 74, in wrapper
retval = func(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/numbers.py", line 2095, in __new__
ival = int(i)
File "/data/users/ezyang/a/pytorch/torch/__init__.py", line 242, in __int__
return self.node.int_()
File "/data/users/ezyang/a/pytorch/torch/fx/experimental/symbolic_shapes.py", line 210, in int_
raise RuntimeError("Trying to extract a concrete int out of a symbolic int")
RuntimeError: Trying to extract a concrete int out of a symbolic int
While executing %primals_214 : [#users=1] = placeholder[target=primals_214]
Original traceback:
None
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 476, in compile_subgraph
self.compile_and_call_fx_graph(tx, list(reversed(stack_values)), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised RuntimeError: Trying to extract a concrete int out of a symbolic int
While executing %primals_214 : [#users=1] = placeholder[target=primals_214]
Original traceback:
None
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 313 ops
cuda train timm_regnet ERROR:common:compile_fx raised RuntimeError: Trying to extract a concrete int out of a symbolic int
While executing %primals_265 : [#users=1] = placeholder[target=primals_265]
Original traceback:
None
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/cache.py", line 70, in wrapper
retval = cfunc(*args, **kwargs)
TypeError: unhashable type: 'SymInt'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 231, in placeholder
sizes, strides = self.static_sizes_strides(example)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 62, in static_sizes_strides
size = [sympy.Integer(i) for i in ex.size()]
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 62, in <listcomp>
size = [sympy.Integer(i) for i in ex.size()]
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/cache.py", line 74, in wrapper
retval = func(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/numbers.py", line 2095, in __new__
ival = int(i)
File "/data/users/ezyang/a/pytorch/torch/__init__.py", line 242, in __int__
return self.node.int_()
File "/data/users/ezyang/a/pytorch/torch/fx/experimental/symbolic_shapes.py", line 210, in int_
raise RuntimeError("Trying to extract a concrete int out of a symbolic int")
RuntimeError: Trying to extract a concrete int out of a symbolic int
While executing %primals_265 : [#users=1] = placeholder[target=primals_265]
Original traceback:
None
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 476, in compile_subgraph
self.compile_and_call_fx_graph(tx, list(reversed(stack_values)), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised RuntimeError: Trying to extract a concrete int out of a symbolic int
While executing %primals_265 : [#users=1] = placeholder[target=primals_265]
Original traceback:
None
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 458 ops
cuda train timm_resnest ERROR:common:name 's0' is not defined
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/timm/models/resnet.py", line 716, in forward
def forward(self, x):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2340, in forward
return compiled_fn(full_args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 887, in g
return f(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1903, in debug_compiled_function
return compiled_function(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1718, in compiled_function
all_outs = CompiledFunction.apply(*args_with_synthetic_bases)
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 419, in apply
return super().apply(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1581, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 912, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 199, in run
return model(new_inputs)
File "/tmp/torchinductor_ezyang/k2/ck2u24qupvv2picybie2m4mekfm6wvvpboaxlvb76jq2jikkuwka.py", line 2178, in call
return (buf1, buf2, buf5, buf6, buf9, buf10, buf15, buf16, buf19, buf20, buf26, buf27, buf34, buf35, buf37, buf38, buf42, buf43, buf46, buf47, buf53, buf54, buf62, buf63, buf66, buf67, buf71, buf72, buf75, buf76, buf82, buf83, buf91, buf92, buf95, buf96, buf100, buf101, buf104, buf105, buf111, buf112, buf120, buf121, buf124, buf125, buf129, primals_1, primals_2, primals_4, primals_5, primals_7, primals_8, primals_10, primals_11, primals_13, primals_14, primals_16, primals_18, primals_20, primals_22, primals_23, primals_25, primals_26, primals_28, primals_29, primals_31, primals_32, primals_34, primals_36, primals_38, primals_40, primals_41, primals_43, primals_44, primals_46, primals_47, primals_49, primals_50, primals_52, primals_54, primals_56, primals_58, primals_59, primals_61, primals_62, primals_64, primals_65, primals_67, primals_68, primals_70, primals_72, primals_74, primals_76, primals_77, primals_79, primals_80, primals_153, buf0, buf1, buf2, buf3, buf4, buf5, buf6, buf7, buf8, buf9, buf10, buf11, buf12, buf13, buf14, buf15, buf16, buf17, buf18, buf19, buf20, buf21, buf23, buf25, buf26, buf27, buf28, buf31, buf33, buf34, buf35, buf36, buf37, buf38, buf40, buf41, buf42, buf43, buf44, buf45, buf46, buf47, buf48, buf50, buf52, buf53, buf54, buf55, buf58, buf60, buf61, buf62, buf63, buf64, buf65, buf66, buf67, buf69, buf70, buf71, buf72, buf73, buf74, buf75, buf76, buf77, buf79, buf81, buf82, buf83, buf84, buf87, buf89, buf90, buf91, buf92, buf93, buf94, buf95, buf96, buf98, buf99, buf100, buf101, buf102, buf103, buf104, buf105, buf106, buf108, buf110, buf111, buf112, buf113, buf116, buf118, buf119, buf120, buf121, buf122, buf123, buf124, buf125, as_strided(buf128, (4, 2048), (2048, 1)), as_strided(primals_82, (1000, 2048), (2048, 1)), buf130, s0, 56, 56, 128, 56, 56, 256, 28, 28, 512, 14, 14, 1024, )
NameError: name 's0' is not defined
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 180 ops
cuda train timm_vision_transformer ERROR:common:name 's0' is not defined
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/timm/models/vision_transformer.py", line 449, in forward
def forward(self, x):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2340, in forward
return compiled_fn(full_args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 887, in g
return f(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1903, in debug_compiled_function
return compiled_function(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1718, in compiled_function
all_outs = CompiledFunction.apply(*args_with_synthetic_bases)
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 419, in apply
return super().apply(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1581, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 912, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 199, in run
return model(new_inputs)
File "/tmp/torchinductor_ezyang/d5/cd5yxyrdmu3enl64j7qamsdiqvonep4pu6qxgfruo7s6dwa3oz7w.py", line 1690, in call
return (buf304, primals_2, primals_3, primals_5, primals_6, primals_11, primals_12, primals_17, primals_18, primals_23, primals_24, primals_29, primals_30, primals_35, primals_36, primals_41, primals_42, primals_47, primals_48, primals_53, primals_54, primals_59, primals_60, primals_65, primals_66, primals_71, primals_72, primals_77, primals_78, primals_83, primals_84, primals_89, primals_90, primals_95, primals_96, primals_101, primals_102, primals_107, primals_108, primals_113, primals_114, primals_119, primals_120, primals_125, primals_126, primals_131, primals_132, primals_137, primals_138, primals_143, primals_144, primals_149, primals_153, buf3, buf7, buf8, as_strided(buf11, (24, 197, 64), (12608, 64, 1)), buf16, as_strided(buf19, (788, 384), (384, 1)), buf24, as_strided(buf26, (4, 197, 1536), (302592, 1536, 1)), buf32, as_strided(buf35, (24, 197, 64), (12608, 64, 1)), buf40, as_strided(buf43, (788, 384), (384, 1)), buf49, as_strided(buf51, (4, 197, 1536), (302592, 1536, 1)), buf57, as_strided(buf60, (24, 197, 64), (12608, 64, 1)), buf65, as_strided(buf68, (788, 384), (384, 1)), buf73, as_strided(buf75, (4, 197, 1536), (302592, 1536, 1)), buf81, as_strided(buf84, (24, 197, 64), (12608, 64, 1)), buf89, as_strided(buf92, (788, 384), (384, 1)), buf98, as_strided(buf100, (4, 197, 1536), (302592, 1536, 1)), buf106, as_strided(buf109, (24, 197, 64), (12608, 64, 1)), buf114, as_strided(buf117, (788, 384), (384, 1)), buf122, as_strided(buf124, (4, 197, 1536), (302592, 1536, 1)), buf130, as_strided(buf133, (24, 197, 64), (12608, 64, 1)), buf138, as_strided(buf141, (788, 384), (384, 1)), buf147, as_strided(buf149, (4, 197, 1536), (302592, 1536, 1)), buf155, as_strided(buf158, (24, 197, 64), (12608, 64, 1)), buf163, as_strided(buf166, (788, 384), (384, 1)), buf171, as_strided(buf173, (4, 197, 1536), (302592, 1536, 1)), buf179, as_strided(buf182, (24, 197, 64), (12608, 64, 1)), buf187, as_strided(buf190, (788, 384), (384, 1)), buf196, as_strided(buf198, (4, 197, 1536), (302592, 1536, 1)), buf204, as_strided(buf207, (24, 197, 64), (12608, 64, 1)), buf212, as_strided(buf215, (788, 384), (384, 1)), buf220, as_strided(buf222, (4, 197, 1536), (302592, 1536, 1)), buf228, as_strided(buf231, (24, 197, 64), (12608, 64, 1)), buf236, as_strided(buf239, (788, 384), (384, 1)), buf245, as_strided(buf247, (4, 197, 1536), (302592, 1536, 1)), buf253, as_strided(buf256, (24, 197, 64), (12608, 64, 1)), buf261, as_strided(buf264, (788, 384), (384, 1)), buf269, as_strided(buf271, (4, 197, 1536), (302592, 1536, 1)), buf277, as_strided(buf280, (24, 197, 64), (12608, 64, 1)), buf285, as_strided(buf288, (788, 384), (384, 1)), buf294, as_strided(buf296, (4, 197, 1536), (302592, 1536, 1)), buf302, as_strided(buf303, (4, 384), (75648, 1)), as_strided(primals_151, (1000, 384), (384, 1)), buf305, as_strided(primals_147, (384, 1536), (1536, 1)), as_strided(primals_145, (1536, 384), (384, 1)), buf306, as_strided(primals_141, (384, 384), (384, 1)), as_strided(buf286, (24, 64, 197), (12608, 1, 64)), as_strided(buf281, (24, 197, 64), (12608, 1, 197)), as_strided(primals_139, (1152, 384), (384, 1)), buf307, as_strided(primals_135, (384, 1536), (1536, 1)), as_strided(primals_133, (1536, 384), (384, 1)), buf308, as_strided(primals_129, (384, 384), (384, 1)), as_strided(buf262, (24, 64, 197), (12608, 1, 64)), as_strided(buf257, (24, 197, 64), (12608, 1, 197)), as_strided(primals_127, (1152, 384), (384, 1)), buf309, as_strided(primals_123, (384, 1536), (1536, 1)), as_strided(primals_121, (1536, 384), (384, 1)), buf310, as_strided(primals_117, (384, 384), (384, 1)), as_strided(buf237, (24, 64, 197), (12608, 1, 64)), as_strided(buf232, (24, 197, 64), (12608, 1, 197)), as_strided(primals_115, (1152, 384), (384, 1)), buf311, as_strided(primals_111, (384, 1536), (1536, 1)), as_strided(primals_109, (1536, 384), (384, 1)), buf312, as_strided(primals_105, (384, 384), (384, 1)), as_strided(buf213, (24, 64, 197), (12608, 1, 64)), as_strided(buf208, (24, 197, 64), (12608, 1, 197)), as_strided(primals_103, (1152, 384), (384, 1)), buf313, as_strided(primals_99, (384, 1536), (1536, 1)), as_strided(primals_97, (1536, 384), (384, 1)), buf314, as_strided(primals_93, (384, 384), (384, 1)), as_strided(buf188, (24, 64, 197), (12608, 1, 64)), as_strided(buf183, (24, 197, 64), (12608, 1, 197)), as_strided(primals_91, (1152, 384), (384, 1)), buf315, as_strided(primals_87, (384, 1536), (1536, 1)), as_strided(primals_85, (1536, 384), (384, 1)), buf316, as_strided(primals_81, (384, 384), (384, 1)), as_strided(buf164, (24, 64, 197), (12608, 1, 64)), as_strided(buf159, (24, 197, 64), (12608, 1, 197)), as_strided(primals_79, (1152, 384), (384, 1)), buf317, as_strided(primals_75, (384, 1536), (1536, 1)), as_strided(primals_73, (1536, 384), (384, 1)), buf318, as_strided(primals_69, (384, 384), (384, 1)), as_strided(buf139, (24, 64, 197), (12608, 1, 64)), as_strided(buf134, (24, 197, 64), (12608, 1, 197)), as_strided(primals_67, (1152, 384), (384, 1)), buf319, as_strided(primals_63, (384, 1536), (1536, 1)), as_strided(primals_61, (1536, 384), (384, 1)), buf320, as_strided(primals_57, (384, 384), (384, 1)), as_strided(buf115, (24, 64, 197), (12608, 1, 64)), as_strided(buf110, (24, 197, 64), (12608, 1, 197)), as_strided(primals_55, (1152, 384), (384, 1)), buf321, as_strided(primals_51, (384, 1536), (1536, 1)), as_strided(primals_49, (1536, 384), (384, 1)), buf322, as_strided(primals_45, (384, 384), (384, 1)), as_strided(buf90, (24, 64, 197), (12608, 1, 64)), as_strided(buf85, (24, 197, 64), (12608, 1, 197)), as_strided(primals_43, (1152, 384), (384, 1)), buf323, as_strided(primals_39, (384, 1536), (1536, 1)), as_strided(primals_37, (1536, 384), (384, 1)), buf324, as_strided(primals_33, (384, 384), (384, 1)), as_strided(buf66, (24, 64, 197), (12608, 1, 64)), as_strided(buf61, (24, 197, 64), (12608, 1, 197)), as_strided(primals_31, (1152, 384), (384, 1)), buf325, as_strided(primals_27, (384, 1536), (1536, 1)), as_strided(primals_25, (1536, 384), (384, 1)), buf326, as_strided(primals_21, (384, 384), (384, 1)), as_strided(buf41, (24, 64, 197), (12608, 1, 64)), as_strided(buf36, (24, 197, 64), (12608, 1, 197)), as_strided(primals_19, (1152, 384), (384, 1)), buf327, as_strided(primals_15, (384, 1536), (1536, 1)), as_strided(primals_13, (1536, 384), (384, 1)), buf328, as_strided(primals_9, (384, 384), (384, 1)), as_strided(buf17, (24, 64, 197), (12608, 1, 64)), as_strided(buf12, (24, 197, 64), (12608, 1, 197)), as_strided(primals_7, (1152, 384), (384, 1)), s0, 14, 14, 197, 197*s0, 197, 6, 6*s0, 384, 197, 197*s0, 197, 6, 6*s0, 384, 197, 197*s0, 197, 6, 6*s0, 384, 197, 197*s0, 197, 6, 6*s0, 384, 197, 197*s0, 197, 6, 6*s0, 384, 197, 197*s0, 197, 6, 6*s0, 384, 197, 197*s0, 197, 6, 6*s0, 384, 197, 197*s0, 197, 6, 6*s0, 384, 197, 197*s0, 197, 6, 6*s0, 384, 197, 197*s0, 197, 6, 6*s0, 384, 197, 197*s0, 197, 6, 6*s0, 384, 197, 197*s0, 197, 6, 6*s0, 384, 197, 197*s0, s0, 6, 6, 6, 6, 6, 6, 6, 6, 6, 6, 6, 6, 196, )
NameError: name 's0' is not defined
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 441 ops
cuda train timm_vision_transformer_large PASS
Dynamo produced 0 graph(s) covering 0 ops
cuda train timm_vovnet ERROR:common:compile_fx raised RuntimeError: Trying to extract a concrete int out of a symbolic int
While executing %primals_120 : [#users=1] = placeholder[target=primals_120]
Original traceback:
None
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/cache.py", line 70, in wrapper
retval = cfunc(*args, **kwargs)
TypeError: unhashable type: 'SymInt'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 231, in placeholder
sizes, strides = self.static_sizes_strides(example)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 62, in static_sizes_strides
size = [sympy.Integer(i) for i in ex.size()]
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 62, in <listcomp>
size = [sympy.Integer(i) for i in ex.size()]
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/cache.py", line 74, in wrapper
retval = func(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/numbers.py", line 2095, in __new__
ival = int(i)
File "/data/users/ezyang/a/pytorch/torch/__init__.py", line 242, in __int__
return self.node.int_()
File "/data/users/ezyang/a/pytorch/torch/fx/experimental/symbolic_shapes.py", line 210, in int_
raise RuntimeError("Trying to extract a concrete int out of a symbolic int")
RuntimeError: Trying to extract a concrete int out of a symbolic int
While executing %primals_120 : [#users=1] = placeholder[target=primals_120]
Original traceback:
None
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 476, in compile_subgraph
self.compile_and_call_fx_graph(tx, list(reversed(stack_values)), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised RuntimeError: Trying to extract a concrete int out of a symbolic int
While executing %primals_120 : [#users=1] = placeholder[target=primals_120]
Original traceback:
None
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 169 ops
cuda train tts_angular ERROR:common:
from user code:
File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/models/tts_angular/model.py", line 18, in <graph break in forward>
o, (_, _) = self.lstm(x)
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 1092, in run_node
return nnmodule(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/rnn.py", line 776, in forward
result = _VF.lstm(input, hx, self._flat_weights, self.bias, self.num_layers,
RuntimeError: Cannot call sizes() on tensor with symbolic sizes/strides
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 1046, in get_fake_value
return wrap_fake_exception(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 712, in wrap_fake_exception
return fn()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 1047, in <lambda>
lambda: run_node(tx.output, node, args, kwargs, nnmodule)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 1096, in run_node
raise RuntimeError(
RuntimeError: Failed running call_module self_lstm(*(FakeTensor(FakeTensor(..., device='meta', size=(s0, s1, 40)), cuda:0),), **{}):
Cannot call sizes() on tensor with symbolic sizes/strides
(scroll up for backtrace)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/models/tts_angular/model.py", line 59, in forward
d = self.layers(x)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/container.py", line 204, in forward
input = module(input)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/models/tts_angular/model.py", line 17, in forward
self.lstm.flatten_parameters()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 305, in wrapper
return inner_fn(self, inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 956, in CALL_FUNCTION
self.call_function(fn, args, {})
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 430, in call_function
self.push(fn.call_function(self, args, kwargs))
File "/data/users/ezyang/a/pytorch/torch/_dynamo/variables/nn_module.py", line 201, in call_function
return wrap_fx_proxy(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/variables/builder.py", line 729, in wrap_fx_proxy
return wrap_fx_proxy_cls(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/variables/builder.py", line 766, in wrap_fx_proxy_cls
example_value = get_fake_value(proxy.node, tx)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 1066, in get_fake_value
raise TorchRuntimeError() from e
torch._dynamo.exc.TorchRuntimeError:
from user code:
File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/models/tts_angular/model.py", line 18, in <graph break in forward>
o, (_, _) = self.lstm(x)
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 0 ops
cuda train vgg16 ERROR:common:name 's0' is not defined
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/torchvision/torchvision/models/vgg.py", line 65, in forward
def forward(self, x: torch.Tensor) -> torch.Tensor:
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2340, in forward
return compiled_fn(full_args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 887, in g
return f(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1903, in debug_compiled_function
return compiled_function(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1718, in compiled_function
all_outs = CompiledFunction.apply(*args_with_synthetic_bases)
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 419, in apply
return super().apply(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1581, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 912, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 199, in run
return model(new_inputs)
File "/tmp/torchinductor_ezyang/4b/c4bucu2bcj7r2kirrua5nbo4simoi6tw43w5hasphd5ltxvo7gjj.py", line 461, in call
return (buf36, primals_1, primals_3, primals_5, primals_7, primals_9, primals_11, primals_13, primals_15, primals_17, primals_19, primals_21, primals_23, primals_25, primals_33, buf1, buf3, buf4, buf37, buf6, buf8, buf9, buf38, buf11, buf13, buf15, buf16, buf39, buf18, buf20, buf22, buf23, buf40, buf25, buf27, buf29, buf30, buf41, as_strided(buf31, (4, 25088), (25088, 1)), buf33, buf35, as_strided(primals_31, (1000, 4096), (4096, 1)), as_strided(primals_29, (4096, 4096), (4096, 1)), as_strided(primals_27, (4096, 25088), (25088, 1)), s0, )
NameError: name 's0' is not defined
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 40 ops
cuda train vision_maskrcnn Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 373, in <module>
main(TorchBenchmarkRunner(), original_dir)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1670, in main
return maybe_fresh_cache(run, args.cold_start_latency and args.only)(
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 842, in inner
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 2044, in run
runner.run_one_model(
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1333, in run_one_model
status = self.check_accuracy(
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1139, in check_accuracy
if not same(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 771, in same
return len(ref) == len(res) and all(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 772, in <genexpr>
same(ai, bi, fp64_refi, cos_similarity, tol, equal_nan, exact_dtype)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 771, in same
return len(ref) == len(res) and all(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 772, in <genexpr>
same(ai, bi, fp64_refi, cos_similarity, tol, equal_nan, exact_dtype)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 782, in same
same(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 839, in same
ref_error = rmse(fp64_ref, ref).item()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 754, in rmse
return torch.sqrt(torch.mean(torch.square(ref - res)))
RuntimeError: The size of tensor a (28) must match the size of tensor b (29) at non-singleton dimension 0
cuda train yolov3 [2022-12-12 07:03:14,612] torch._inductor.graph: [WARNING] Creating implicit fallback for:
target: <built-in function truediv>
args[0]: 25165824
args[1]: 32
ERROR:common:compile_fx raised LoweringException: TypeError: truediv expected 2 arguments, got 0
target: <built-in function truediv>
args[0]: 25165824
args[1]: 32
While executing %truediv : [#users=2] = call_function[target=operator.truediv](args = (%sym_numel, 32), kwargs = {})
Original traceback:
None
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 272, in call_function
out = lowerings[target](*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 221, in wrapped
return decomp_fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 991, in handler
TensorBox.create, ir.FallbackKernel.create(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 3011, in create
) = cls.process_kernel(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 2406, in process_kernel
example_output = kernel(*new_args, **new_kwargs)
TypeError: truediv expected 2 arguments, got 0
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 275, in call_function
raise LoweringException(e, target, args, kwargs) from e
torch._inductor.exc.LoweringException: TypeError: truediv expected 2 arguments, got 0
target: <built-in function truediv>
args[0]: 25165824
args[1]: 32
While executing %truediv : [#users=2] = call_function[target=operator.truediv](args = (%sym_numel, 32), kwargs = {})
Original traceback:
None
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 356, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 357, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/torchbench.py", line 359, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/models/yolov3/yolo_models.py", line 238, in forward
return self.forward_once(x)
File "/data/users/ezyang/a/torchbenchmark/torchbenchmark/models/yolov3/yolo_models.py", line 292, in forward_once
x = module(x)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 476, in compile_subgraph
self.compile_and_call_fx_graph(tx, list(reversed(stack_values)), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised LoweringException: TypeError: truediv expected 2 arguments, got 0
target: <built-in function truediv>
args[0]: 25165824
args[1]: 32
While executing %truediv : [#users=2] = call_function[target=operator.truediv](args = (%sym_numel, 32), kwargs = {})
Original traceback:
None
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 3 ops
ERROR
cuda train AlbertForMaskedLM ERROR:common:compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 440, in codegen
self.scheduler.codegen()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/scheduler.py", line 1129, in codegen
self.get_backend(device).codegen_nodes(node.get_nodes())
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1284, in codegen_nodes
return self.codegen_node_schedule(node_schedule, numel, rnumel)
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1336, in codegen_node_schedule
src_code = kernel.codegen_kernel()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1035, in codegen_kernel
"signature": dict(enumerate(map(signature_of, signature))),
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 57, in signature_of
return JITFunction._key_of(V.graph.sizevars.size_hint(arg.expr))
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 367, in size_hint
return int(out)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 990, in forward
outputs = self.albert(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 737, in forward
encoder_outputs = self.encoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 2 graph(s) covering 560 ops
cuda train AlbertForQuestionAnswering ERROR:common:compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 440, in codegen
self.scheduler.codegen()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/scheduler.py", line 1129, in codegen
self.get_backend(device).codegen_nodes(node.get_nodes())
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1284, in codegen_nodes
return self.codegen_node_schedule(node_schedule, numel, rnumel)
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1336, in codegen_node_schedule
src_code = kernel.codegen_kernel()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1035, in codegen_kernel
"signature": dict(enumerate(map(signature_of, signature))),
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 57, in signature_of
return JITFunction._key_of(V.graph.sizevars.size_hint(arg.expr))
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 367, in size_hint
return int(out)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 1274, in forward
outputs = self.albert(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 737, in forward
encoder_outputs = self.encoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 2 graph(s) covering 560 ops
cuda train AllenaiLongformerBase [2022-12-12 07:05:06,965] torch._inductor.ir: [WARNING] Using FallbackKernel: aten.cumsum
ERROR:common:compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 952, in aot_dispatch_base
compiled_fw = aot_config.fw_compiler(fw_module, flat_args)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 437, in codegen
self.init_wrapper_code()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 431, in init_wrapper_code
self.wrapper_code = WrapperCodeGen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 302, in __init__
self.write_prefix()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 335, in write_prefix
V.graph.sizevars.codegen(self.wrapper_call, V.graph.graph_inputs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 494, in codegen
assert not needed
AssertionError
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1813, in forward
outputs = self.longformer(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1696, in forward
padding_len, input_ids, attention_mask, token_type_ids, position_ids, inputs_embeds = self._pad_to_window_size(
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1715, in <graph break in forward>
encoder_outputs = self.encoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 332, in wrapper
self.output.compile_subgraph(self, reason=reason)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 2 graph(s) covering 30 ops
cuda train BartForCausalLM [2022-12-12 07:05:26,654] torch._inductor.ir: [WARNING] DeviceCopy
ERROR:common:compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 437, in codegen
self.init_wrapper_code()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 431, in init_wrapper_code
self.wrapper_code = WrapperCodeGen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 302, in __init__
self.write_prefix()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 335, in write_prefix
V.graph.sizevars.codegen(self.wrapper_call, V.graph.graph_inputs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 494, in codegen
assert not needed
AssertionError
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1839, in forward
outputs = self.model.decoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1098, in forward
layer_outputs = decoder_layer(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 418, in forward
hidden_states, self_attn_weights, present_key_value = self.self_attn(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 3 graph(s) covering 49 ops
cuda train BartForConditionalGeneration ERROR:common:compile_fx raised LoweringException: TypeError: Argument of Integer should be of numeric type, got s0.
target: aten.new_zeros.default
args[0]: TensorBox(StorageBox(
InputBuffer(name='arg0_1', layout=FixedLayout('cuda', torch.int64, size=[1, s0], stride=[s0, 1]))
))
args[1]: [1, s0]
kwargs: {'dtype': torch.int64, 'layout': torch.strided, 'device': device(type='cuda', index=0), 'pin_memory': False}
While executing %new_zeros : [#users=5] = call_function[target=torch.ops.aten.new_zeros.default](args = (%arg0_1, [%sym_size, %sym_size_1]), kwargs = {dtype: torch.int64, layout: torch.strided, device: cuda:0, pin_memory: False})
Original traceback:
Module stack: {}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 79, in shift_tokens_right
shifted_input_ids = input_ids.new_zeros(input_ids.shape)
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/numbers.py", line 2095, in __new__
ival = int(i)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 272, in call_function
out = lowerings[target](*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 221, in wrapped
return decomp_fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 1672, in _new_constant
size = [sympy.Integer(s) for s in size]
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 1672, in <listcomp>
size = [sympy.Integer(s) for s in size]
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/cache.py", line 70, in wrapper
retval = cfunc(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/numbers.py", line 2097, in __new__
raise TypeError(
TypeError: Argument of Integer should be of numeric type, got s0.
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 952, in aot_dispatch_base
compiled_fw = aot_config.fw_compiler(fw_module, flat_args)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 275, in call_function
raise LoweringException(e, target, args, kwargs) from e
torch._inductor.exc.LoweringException: TypeError: Argument of Integer should be of numeric type, got s0.
target: aten.new_zeros.default
args[0]: TensorBox(StorageBox(
InputBuffer(name='arg0_1', layout=FixedLayout('cuda', torch.int64, size=[1, s0], stride=[s0, 1]))
))
args[1]: [1, s0]
kwargs: {'dtype': torch.int64, 'layout': torch.strided, 'device': device(type='cuda', index=0), 'pin_memory': False}
While executing %new_zeros : [#users=5] = call_function[target=torch.ops.aten.new_zeros.default](args = (%arg0_1, [%sym_size, %sym_size_1]), kwargs = {dtype: torch.int64, layout: torch.strided, device: cuda:0, pin_memory: False})
Original traceback:
Module stack: {}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 79, in shift_tokens_right
shifted_input_ids = input_ids.new_zeros(input_ids.shape)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1349, in forward
decoder_input_ids = shift_tokens_right(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 476, in compile_subgraph
self.compile_and_call_fx_graph(tx, list(reversed(stack_values)), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised LoweringException: TypeError: Argument of Integer should be of numeric type, got s0.
target: aten.new_zeros.default
args[0]: TensorBox(StorageBox(
InputBuffer(name='arg0_1', layout=FixedLayout('cuda', torch.int64, size=[1, s0], stride=[s0, 1]))
))
args[1]: [1, s0]
kwargs: {'dtype': torch.int64, 'layout': torch.strided, 'device': device(type='cuda', index=0), 'pin_memory': False}
While executing %new_zeros : [#users=5] = call_function[target=torch.ops.aten.new_zeros.default](args = (%arg0_1, [%sym_size, %sym_size_1]), kwargs = {dtype: torch.int64, layout: torch.strided, device: cuda:0, pin_memory: False})
Original traceback:
Module stack: {}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 79, in shift_tokens_right
shifted_input_ids = input_ids.new_zeros(input_ids.shape)
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 8 ops
cuda train BertForMaskedLM ERROR:common:compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 440, in codegen
self.scheduler.codegen()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/scheduler.py", line 1129, in codegen
self.get_backend(device).codegen_nodes(node.get_nodes())
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1284, in codegen_nodes
return self.codegen_node_schedule(node_schedule, numel, rnumel)
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1336, in codegen_node_schedule
src_code = kernel.codegen_kernel()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1035, in codegen_kernel
"signature": dict(enumerate(map(signature_of, signature))),
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 57, in signature_of
return JITFunction._key_of(V.graph.sizevars.size_hint(arg.expr))
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 367, in size_hint
return int(out)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1351, in forward
outputs = self.bert(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1018, in forward
encoder_outputs = self.encoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 3 graph(s) covering 552 ops
cuda train BertForQuestionAnswering ERROR:common:compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 440, in codegen
self.scheduler.codegen()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/scheduler.py", line 1129, in codegen
self.get_backend(device).codegen_nodes(node.get_nodes())
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1284, in codegen_nodes
return self.codegen_node_schedule(node_schedule, numel, rnumel)
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1336, in codegen_node_schedule
src_code = kernel.codegen_kernel()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1035, in codegen_kernel
"signature": dict(enumerate(map(signature_of, signature))),
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 57, in signature_of
return JITFunction._key_of(V.graph.sizevars.size_hint(arg.expr))
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 367, in size_hint
return int(out)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1847, in forward
outputs = self.bert(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1018, in forward
encoder_outputs = self.encoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 3 graph(s) covering 552 ops
cuda train BlenderbotForCausalLM PASS
Dynamo produced 0 graph(s) covering 0 ops
cuda train BlenderbotSmallForCausalLM [2022-12-12 07:08:41,319] torch._inductor.ir: [WARNING] DeviceCopy
ERROR:common:compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 440, in codegen
self.scheduler.codegen()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/scheduler.py", line 1129, in codegen
self.get_backend(device).codegen_nodes(node.get_nodes())
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1284, in codegen_nodes
return self.codegen_node_schedule(node_schedule, numel, rnumel)
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1336, in codegen_node_schedule
src_code = kernel.codegen_kernel()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1035, in codegen_kernel
"signature": dict(enumerate(map(signature_of, signature))),
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 57, in signature_of
return JITFunction._key_of(V.graph.sizevars.size_hint(arg.expr))
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 367, in size_hint
return int(out)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/blenderbot_small/modeling_blenderbot_small.py", line 1529, in forward
outputs = self.model.decoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/blenderbot_small/modeling_blenderbot_small.py", line 1033, in forward
layer_outputs = decoder_layer(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/blenderbot_small/modeling_blenderbot_small.py", line 407, in forward
hidden_states, self_attn_weights, present_key_value = self.self_attn(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 516, in step
self.output.compile_subgraph(self, partial_convert=True)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 2 graph(s) covering 38 ops
cuda train BlenderbotSmallForConditionalGeneration ERROR:common:compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 440, in codegen
self.scheduler.codegen()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/scheduler.py", line 1129, in codegen
self.get_backend(device).codegen_nodes(node.get_nodes())
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1284, in codegen_nodes
return self.codegen_node_schedule(node_schedule, numel, rnumel)
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1336, in codegen_node_schedule
src_code = kernel.codegen_kernel()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1035, in codegen_kernel
"signature": dict(enumerate(map(signature_of, signature))),
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 57, in signature_of
return JITFunction._key_of(V.graph.sizevars.size_hint(arg.expr))
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 367, in size_hint
return int(out)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/blenderbot_small/modeling_blenderbot_small.py", line 1292, in forward
outputs = self.model(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/blenderbot_small/modeling_blenderbot_small.py", line 1155, in forward
encoder_outputs = self.encoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/blenderbot_small/modeling_blenderbot_small.py", line 780, in forward
layer_outputs = encoder_layer(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/blenderbot_small/modeling_blenderbot_small.py", line 311, in forward
hidden_states, attn_weights, _ = self.self_attn(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 516, in step
self.output.compile_subgraph(self, partial_convert=True)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 25 ops
cuda train CamemBert ERROR:common:compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 952, in aot_dispatch_base
compiled_fw = aot_config.fw_compiler(fw_module, flat_args)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 437, in codegen
self.init_wrapper_code()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 431, in init_wrapper_code
self.wrapper_code = WrapperCodeGen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 302, in __init__
self.write_prefix()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 335, in write_prefix
V.graph.sizevars.codegen(self.wrapper_call, V.graph.graph_inputs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 494, in codegen
assert not needed
AssertionError
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/roberta/modeling_roberta.py", line 1095, in forward
outputs = self.roberta(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 516, in step
self.output.compile_subgraph(self, partial_convert=True)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 5 ops
cuda train DebertaForMaskedLM ERROR:common:compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 952, in aot_dispatch_base
compiled_fw = aot_config.fw_compiler(fw_module, flat_args)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 440, in codegen
self.scheduler.codegen()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/scheduler.py", line 1129, in codegen
self.get_backend(device).codegen_nodes(node.get_nodes())
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1284, in codegen_nodes
return self.codegen_node_schedule(node_schedule, numel, rnumel)
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1336, in codegen_node_schedule
src_code = kernel.codegen_kernel()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1035, in codegen_kernel
"signature": dict(enumerate(map(signature_of, signature))),
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 57, in signature_of
return JITFunction._key_of(V.graph.sizevars.size_hint(arg.expr))
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 367, in size_hint
return int(out)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/deberta/modeling_deberta.py", line 1041, in forward
outputs = self.deberta(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/deberta/modeling_deberta.py", line 946, in forward
embedding_output = self.embeddings(
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/deberta/modeling_deberta.py", line 954, in <graph break in forward>
encoder_outputs = self.encoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/deberta/modeling_deberta.py", line 414, in forward
attention_mask = self.get_attention_mask(attention_mask)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 476, in compile_subgraph
self.compile_and_call_fx_graph(tx, list(reversed(stack_values)), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 3 graph(s) covering 24 ops
cuda train DebertaForQuestionAnswering ERROR:common:compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 952, in aot_dispatch_base
compiled_fw = aot_config.fw_compiler(fw_module, flat_args)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 440, in codegen
self.scheduler.codegen()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/scheduler.py", line 1129, in codegen
self.get_backend(device).codegen_nodes(node.get_nodes())
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1284, in codegen_nodes
return self.codegen_node_schedule(node_schedule, numel, rnumel)
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1336, in codegen_node_schedule
src_code = kernel.codegen_kernel()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1035, in codegen_kernel
"signature": dict(enumerate(map(signature_of, signature))),
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 57, in signature_of
return JITFunction._key_of(V.graph.sizevars.size_hint(arg.expr))
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 367, in size_hint
return int(out)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/deberta/modeling_deberta.py", line 1369, in forward
outputs = self.deberta(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/deberta/modeling_deberta.py", line 946, in forward
embedding_output = self.embeddings(
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/deberta/modeling_deberta.py", line 954, in <graph break in forward>
encoder_outputs = self.encoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/deberta/modeling_deberta.py", line 414, in forward
attention_mask = self.get_attention_mask(attention_mask)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 476, in compile_subgraph
self.compile_and_call_fx_graph(tx, list(reversed(stack_values)), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 3 graph(s) covering 24 ops
cuda train DebertaV2ForMaskedLM PASS
Dynamo produced 0 graph(s) covering 0 ops
cuda train DebertaV2ForQuestionAnswering ERROR:common:compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 952, in aot_dispatch_base
compiled_fw = aot_config.fw_compiler(fw_module, flat_args)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 440, in codegen
self.scheduler.codegen()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/scheduler.py", line 1129, in codegen
self.get_backend(device).codegen_nodes(node.get_nodes())
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1284, in codegen_nodes
return self.codegen_node_schedule(node_schedule, numel, rnumel)
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1336, in codegen_node_schedule
src_code = kernel.codegen_kernel()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1035, in codegen_kernel
"signature": dict(enumerate(map(signature_of, signature))),
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 57, in signature_of
return JITFunction._key_of(V.graph.sizevars.size_hint(arg.expr))
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 367, in size_hint
return int(out)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/deberta_v2/modeling_deberta_v2.py", line 1469, in forward
outputs = self.deberta(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/deberta_v2/modeling_deberta_v2.py", line 1042, in forward
embedding_output = self.embeddings(
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/deberta_v2/modeling_deberta_v2.py", line 1050, in <graph break in forward>
encoder_outputs = self.encoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/deberta_v2/modeling_deberta_v2.py", line 465, in forward
attention_mask = self.get_attention_mask(attention_mask)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 476, in compile_subgraph
self.compile_and_call_fx_graph(tx, list(reversed(stack_values)), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 2 graph(s) covering 12 ops
WARNING:__main__:Sequence Length not defined for DistilBertForMaskedLM. Choosing 128 arbitrarily
cuda train DistilBertForMaskedLM ERROR:common:compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 952, in aot_dispatch_base
compiled_fw = aot_config.fw_compiler(fw_module, flat_args)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 437, in codegen
self.init_wrapper_code()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 431, in init_wrapper_code
self.wrapper_code = WrapperCodeGen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 302, in __init__
self.write_prefix()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 335, in write_prefix
V.graph.sizevars.codegen(self.wrapper_call, V.graph.graph_inputs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 494, in codegen
assert not needed
AssertionError
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 649, in forward
dlbrt_output = self.distilbert(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 332, in wrapper
self.output.compile_subgraph(self, reason=reason)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 2 ops
WARNING:__main__:Sequence Length not defined for DistilBertForQuestionAnswering. Choosing 128 arbitrarily
cuda train DistilBertForQuestionAnswering ERROR:common:compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 952, in aot_dispatch_base
compiled_fw = aot_config.fw_compiler(fw_module, flat_args)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 437, in codegen
self.init_wrapper_code()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 431, in init_wrapper_code
self.wrapper_code = WrapperCodeGen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 302, in __init__
self.write_prefix()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 335, in write_prefix
V.graph.sizevars.codegen(self.wrapper_call, V.graph.graph_inputs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 494, in codegen
assert not needed
AssertionError
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 868, in forward
distilbert_output = self.distilbert(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 332, in wrapper
self.output.compile_subgraph(self, reason=reason)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 2 ops
cuda train DistillGPT2 ERROR:common:compile_fx raised AssertionError: s1 is needed but not added
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 437, in codegen
self.init_wrapper_code()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 431, in init_wrapper_code
self.wrapper_code = WrapperCodeGen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 302, in __init__
self.write_prefix()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 335, in write_prefix
V.graph.sizevars.codegen(self.wrapper_call, V.graph.graph_inputs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 493, in codegen
assert shape in added, f"{shape} is needed but not added"
AssertionError: s1 is needed but not added
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1048, in forward
transformer_outputs = self.transformer(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 891, in forward
outputs = block(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 391, in forward
attn_outputs = self.attn(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 332, in forward
attn_output, attn_weights = self._attn(query, key, value, attention_mask, head_mask)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 516, in step
self.output.compile_subgraph(self, partial_convert=True)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 476, in compile_subgraph
self.compile_and_call_fx_graph(tx, list(reversed(stack_values)), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised AssertionError: s1 is needed but not added
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 2 graph(s) covering 33 ops
If you want to use `ElectraForCausalLM` as a standalone, add `is_decoder=True.`
cuda train ElectraForCausalLM ERROR:common:compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 440, in codegen
self.scheduler.codegen()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/scheduler.py", line 1129, in codegen
self.get_backend(device).codegen_nodes(node.get_nodes())
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1284, in codegen_nodes
return self.codegen_node_schedule(node_schedule, numel, rnumel)
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1336, in codegen_node_schedule
src_code = kernel.codegen_kernel()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1035, in codegen_kernel
"signature": dict(enumerate(map(signature_of, signature))),
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 57, in signature_of
return JITFunction._key_of(V.graph.sizevars.size_hint(arg.expr))
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 367, in size_hint
return int(out)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/electra/modeling_electra.py", line 1621, in forward
outputs = self.electra(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/electra/modeling_electra.py", line 917, in forward
hidden_states = self.encoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 3 graph(s) covering 551 ops
cuda train ElectraForQuestionAnswering ERROR:common:compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 440, in codegen
self.scheduler.codegen()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/scheduler.py", line 1129, in codegen
self.get_backend(device).codegen_nodes(node.get_nodes())
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1284, in codegen_nodes
return self.codegen_node_schedule(node_schedule, numel, rnumel)
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1336, in codegen_node_schedule
src_code = kernel.codegen_kernel()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1035, in codegen_kernel
"signature": dict(enumerate(map(signature_of, signature))),
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 57, in signature_of
return JITFunction._key_of(V.graph.sizevars.size_hint(arg.expr))
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 367, in size_hint
return int(out)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/electra/modeling_electra.py", line 1390, in forward
discriminator_hidden_states = self.electra(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/electra/modeling_electra.py", line 917, in forward
hidden_states = self.encoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 3 graph(s) covering 551 ops
cuda train GPT2ForSequenceClassification ERROR:common:compile_fx raised AssertionError: s1 is needed but not added
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 437, in codegen
self.init_wrapper_code()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 431, in init_wrapper_code
self.wrapper_code = WrapperCodeGen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 302, in __init__
self.write_prefix()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 335, in write_prefix
V.graph.sizevars.codegen(self.wrapper_call, V.graph.graph_inputs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 493, in codegen
assert shape in added, f"{shape} is needed but not added"
AssertionError: s1 is needed but not added
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1378, in forward
transformer_outputs = self.transformer(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 891, in forward
outputs = block(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 391, in forward
attn_outputs = self.attn(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 332, in forward
attn_output, attn_weights = self._attn(query, key, value, attention_mask, head_mask)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 516, in step
self.output.compile_subgraph(self, partial_convert=True)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 476, in compile_subgraph
self.compile_and_call_fx_graph(tx, list(reversed(stack_values)), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised AssertionError: s1 is needed but not added
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 2 graph(s) covering 33 ops
cuda train GoogleFnet ERROR:common:compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 440, in codegen
self.scheduler.codegen()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/scheduler.py", line 1129, in codegen
self.get_backend(device).codegen_nodes(node.get_nodes())
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1284, in codegen_nodes
return self.codegen_node_schedule(node_schedule, numel, rnumel)
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1336, in codegen_node_schedule
src_code = kernel.codegen_kernel()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1035, in codegen_kernel
"signature": dict(enumerate(map(signature_of, signature))),
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 57, in signature_of
return JITFunction._key_of(V.graph.sizevars.size_hint(arg.expr))
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 367, in size_hint
return int(out)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/fnet/modeling_fnet.py", line 763, in forward
outputs = self.fnet(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/fnet/modeling_fnet.py", line 604, in forward
encoder_outputs = self.encoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/fnet/modeling_fnet.py", line 308, in forward
layer_outputs = layer_module(hidden_states)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/fnet/modeling_fnet.py", line 267, in forward
self_fourier_outputs = self.fourier(hidden_states)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/fnet/modeling_fnet.py", line 220, in forward
self_outputs = self.self(hidden_states)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 2 graph(s) covering 8 ops
cuda train LayoutLMForMaskedLM ERROR:common:compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 952, in aot_dispatch_base
compiled_fw = aot_config.fw_compiler(fw_module, flat_args)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 437, in codegen
self.init_wrapper_code()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 431, in init_wrapper_code
self.wrapper_code = WrapperCodeGen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 302, in __init__
self.write_prefix()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 335, in write_prefix
V.graph.sizevars.codegen(self.wrapper_call, V.graph.graph_inputs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 494, in codegen
assert not needed
AssertionError
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/layoutlm/modeling_layoutlm.py", line 935, in forward
outputs = self.layoutlm(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/layoutlm/modeling_layoutlm.py", line 820, in forward
embedding_output = self.embeddings(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 516, in step
self.output.compile_subgraph(self, partial_convert=True)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 14 ops
cuda train LayoutLMForSequenceClassification ERROR:common:compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 952, in aot_dispatch_base
compiled_fw = aot_config.fw_compiler(fw_module, flat_args)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 437, in codegen
self.init_wrapper_code()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 431, in init_wrapper_code
self.wrapper_code = WrapperCodeGen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 302, in __init__
self.write_prefix()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 335, in write_prefix
V.graph.sizevars.codegen(self.wrapper_call, V.graph.graph_inputs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 494, in codegen
assert not needed
AssertionError
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/layoutlm/modeling_layoutlm.py", line 1057, in forward
outputs = self.layoutlm(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/layoutlm/modeling_layoutlm.py", line 820, in forward
embedding_output = self.embeddings(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 516, in step
self.output.compile_subgraph(self, partial_convert=True)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 14 ops
WARNING:__main__:Sequence Length not defined for M2M100ForConditionalGeneration. Choosing 128 arbitrarily
cuda train M2M100ForConditionalGeneration [2022-12-12 07:14:30,771] torch._inductor.graph: [WARNING] Creating implicit fallback for:
target: aten.sym_stride
args[0]: TensorBox(StorageBox(
InputBuffer(name='primals_2', layout=FixedLayout('cuda', torch.int64, size=[1, s0], stride=[s0, 1]))
))
args[1]: 0
[2022-12-12 07:14:30,775] torch._inductor.ir: [WARNING] Using FallbackKernel: aten.sym_stride
ERROR:common:compile_fx raised LoweringException: AssertionError: FallbackKernel output type is not supported
target: aten.sym_stride
args[0]: TensorBox(StorageBox(
InputBuffer(name='primals_2', layout=FixedLayout('cuda', torch.int64, size=[1, s0], stride=[s0, 1]))
))
args[1]: 0
While executing %sym_stride : [#users=1] = call_function[target=torch.ops.aten.sym_stride](args = (%primals_2, 0), kwargs = {})
Original traceback:
None
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 272, in call_function
out = lowerings[target](*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 221, in wrapped
return decomp_fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 991, in handler
TensorBox.create, ir.FallbackKernel.create(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 3048, in create
return generate_output(example_output)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 3045, in generate_output
assert output is None, "FallbackKernel output type is not supported"
AssertionError: FallbackKernel output type is not supported
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 275, in call_function
raise LoweringException(e, target, args, kwargs) from e
torch._inductor.exc.LoweringException: AssertionError: FallbackKernel output type is not supported
target: aten.sym_stride
args[0]: TensorBox(StorageBox(
InputBuffer(name='primals_2', layout=FixedLayout('cuda', torch.int64, size=[1, s0], stride=[s0, 1]))
))
args[1]: 0
While executing %sym_stride : [#users=1] = call_function[target=torch.ops.aten.sym_stride](args = (%primals_2, 0), kwargs = {})
Original traceback:
None
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1317, in forward
outputs = self.model(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1190, in forward
encoder_outputs = self.encoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 332, in wrapper
self.output.compile_subgraph(self, reason=reason)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised LoweringException: AssertionError: FallbackKernel output type is not supported
target: aten.sym_stride
args[0]: TensorBox(StorageBox(
InputBuffer(name='primals_2', layout=FixedLayout('cuda', torch.int64, size=[1, s0], stride=[s0, 1]))
))
args[1]: 0
While executing %sym_stride : [#users=1] = call_function[target=torch.ops.aten.sym_stride](args = (%primals_2, 0), kwargs = {})
Original traceback:
None
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 5 ops
cuda train MBartForCausalLM [2022-12-12 07:14:49,201] torch._inductor.ir: [WARNING] DeviceCopy
ERROR:common:compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 437, in codegen
self.init_wrapper_code()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 431, in init_wrapper_code
self.wrapper_code = WrapperCodeGen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 302, in __init__
self.write_prefix()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 335, in write_prefix
V.graph.sizevars.codegen(self.wrapper_call, V.graph.graph_inputs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 494, in codegen
assert not needed
AssertionError
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/mbart/modeling_mbart.py", line 1836, in forward
outputs = self.model.decoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/mbart/modeling_mbart.py", line 1095, in forward
layer_outputs = decoder_layer(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/mbart/modeling_mbart.py", line 426, in forward
hidden_states, self_attn_weights, present_key_value = self.self_attn(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 4 graph(s) covering 49 ops
cuda train MBartForConditionalGeneration [2022-12-12 07:15:46,834] torch._inductor.ir: [WARNING] DeviceCopy
ERROR:common:compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 437, in codegen
self.init_wrapper_code()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 431, in init_wrapper_code
self.wrapper_code = WrapperCodeGen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 302, in __init__
self.write_prefix()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 335, in write_prefix
V.graph.sizevars.codegen(self.wrapper_call, V.graph.graph_inputs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 494, in codegen
assert not needed
AssertionError
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/mbart/modeling_mbart.py", line 1346, in forward
outputs = self.model(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/mbart/modeling_mbart.py", line 1229, in forward
decoder_outputs = self.decoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/mbart/modeling_mbart.py", line 1095, in forward
layer_outputs = decoder_layer(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/mbart/modeling_mbart.py", line 426, in forward
hidden_states, self_attn_weights, present_key_value = self.self_attn(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 332, in wrapper
self.output.compile_subgraph(self, reason=reason)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 42 graph(s) covering 455 ops
WARNING:__main__:Sequence Length not defined for MT5ForConditionalGeneration. Choosing 128 arbitrarily
cuda train MT5ForConditionalGeneration WARNING:common:fp64 golden ref were not generated for MT5ForConditionalGeneration. Setting accuracy check to cosine
ERROR:common:compile_fx raised LoweringException: RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s0
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s0 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
target: aten.arange.default
args[0]: s0
kwargs: {'dtype': torch.int64, 'device': device(type='cuda', index=0), 'pin_memory': False}
While executing %arange : [#users=2] = call_function[target=torch.ops.aten.arange.default](args = (%sym_size,), kwargs = {dtype: torch.int64, device: cuda:0, pin_memory: False})
Original traceback:
Module stack: {'self_encoder': "<class 'transformers.models.t5.modeling_t5.T5Stack'>", 'self_encoder_block_0': "<class 'transformers.models.t5.modeling_t5.T5Block'>", 'sub0_0': "<class 'transformers.models.t5.modeling_t5.T5LayerSelfAttention'>", 'self_encoder_block_0_layer_0_SelfAttention': "<class 'transformers.models.t5.modeling_t5.T5Attention'>"}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 423, in compute_bias
context_position = torch.arange(query_length, dtype=torch.long, device=device)[:, None]
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 519, in forward
position_bias = self.compute_bias(real_seq_length, key_length, device=scores.device)
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 570, in forward
attention_output = self.SelfAttention(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 664, in forward
self_attention_outputs = self.layer[0](
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1033, in forward
layer_outputs = layer_module(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1601, in forward
encoder_outputs = self.encoder(
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 272, in call_function
out = lowerings[target](*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 221, in wrapped
return decomp_fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 1318, in arange
return fallback_arange(
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 991, in handler
TensorBox.create, ir.FallbackKernel.create(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 3011, in create
) = cls.process_kernel(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 2406, in process_kernel
example_output = kernel(*new_args, **new_kwargs)
File "/data/users/ezyang/a/pytorch/torch/_ops.py", line 500, in __call__
return self._op(*args, **kwargs or {})
File "/data/users/ezyang/a/pytorch/torch/_inductor/overrides.py", line 37, in __torch_function__
return func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_ops.py", line 500, in __call__
return self._op(*args, **kwargs or {})
RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s0
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s0 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 275, in call_function
raise LoweringException(e, target, args, kwargs) from e
torch._inductor.exc.LoweringException: RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s0
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s0 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
target: aten.arange.default
args[0]: s0
kwargs: {'dtype': torch.int64, 'device': device(type='cuda', index=0), 'pin_memory': False}
While executing %arange : [#users=2] = call_function[target=torch.ops.aten.arange.default](args = (%sym_size,), kwargs = {dtype: torch.int64, device: cuda:0, pin_memory: False})
Original traceback:
Module stack: {'self_encoder': "<class 'transformers.models.t5.modeling_t5.T5Stack'>", 'self_encoder_block_0': "<class 'transformers.models.t5.modeling_t5.T5Block'>", 'sub0_0': "<class 'transformers.models.t5.modeling_t5.T5LayerSelfAttention'>", 'self_encoder_block_0_layer_0_SelfAttention': "<class 'transformers.models.t5.modeling_t5.T5Attention'>"}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 423, in compute_bias
context_position = torch.arange(query_length, dtype=torch.long, device=device)[:, None]
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 519, in forward
position_bias = self.compute_bias(real_seq_length, key_length, device=scores.device)
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 570, in forward
attention_output = self.SelfAttention(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 664, in forward
self_attention_outputs = self.layer[0](
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1033, in forward
layer_outputs = layer_module(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1601, in forward
encoder_outputs = self.encoder(
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised LoweringException: RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s0
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s0 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
target: aten.arange.default
args[0]: s0
kwargs: {'dtype': torch.int64, 'device': device(type='cuda', index=0), 'pin_memory': False}
While executing %arange : [#users=2] = call_function[target=torch.ops.aten.arange.default](args = (%sym_size,), kwargs = {dtype: torch.int64, device: cuda:0, pin_memory: False})
Original traceback:
Module stack: {'self_encoder': "<class 'transformers.models.t5.modeling_t5.T5Stack'>", 'self_encoder_block_0': "<class 'transformers.models.t5.modeling_t5.T5Block'>", 'sub0_0': "<class 'transformers.models.t5.modeling_t5.T5LayerSelfAttention'>", 'self_encoder_block_0_layer_0_SelfAttention': "<class 'transformers.models.t5.modeling_t5.T5Attention'>"}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 423, in compute_bias
context_position = torch.arange(query_length, dtype=torch.long, device=device)[:, None]
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 519, in forward
position_bias = self.compute_bias(real_seq_length, key_length, device=scores.device)
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 570, in forward
attention_output = self.SelfAttention(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 664, in forward
self_attention_outputs = self.layer[0](
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1033, in forward
layer_outputs = layer_module(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1601, in forward
encoder_outputs = self.encoder(
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 1282 ops
If you want to use `MegatronBertForCausalLM` as a standalone, add `is_decoder=True.`
cuda train MegatronBertForCausalLM ERROR:common:'int' object has no attribute 'size'
While executing %sym_size : [#users=3] = placeholder[target=sym_size]
Original traceback:
Module stack: {'self_cls': "<class 'transformers.models.megatron_bert.modeling_megatron_bert.MegatronBertOnlyMLMHead'>", 'self_cls_predictions': "<class 'transformers.models.megatron_bert.modeling_megatron_bert.MegatronBertLMPredictionHead'>", 'self_cls_predictions_transform': "<class 'transformers.models.megatron_bert.modeling_megatron_bert.MegatronBertPredictionHeadTransform'>", 'self_cls_predictions_transform_dense': "<class 'torch.nn.modules.linear.Linear'>"}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 638, in forward
hidden_states = self.dense(hidden_states)
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 660, in forward
hidden_states = self.transform(hidden_states)
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 672, in forward
prediction_scores = self.predictions(sequence_output)
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 1218, in <graph break in forward>
prediction_scores = self.cls(sequence_output)
Gradient addition node due to multiple use of tensor around:
Module stack: {'self_encoder': "<class 'transformers.models.megatron_bert.modeling_megatron_bert.MegatronBertEncoder'>", 'self_encoder_layer_0': "<class 'transformers.models.megatron_bert.modeling_megatron_bert.MegatronBertLayer'>", 'self_encoder_layer_0_attention': "<class 'transformers.models.megatron_bert.modeling_megatron_bert.MegatronBertAttention'>", 'self_encoder_layer_0_attention_ln': "<class 'torch.nn.modules.normalization.LayerNorm'>"}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 370, in forward
ln_outputs = self.ln(hidden_states)
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 443, in forward
self_attention_outputs = self.attention(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 562, in forward
layer_outputs = layer_module(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 976, in <graph break in forward>
encoder_outputs = self.encoder(
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 485, in <graph break in forward_and_backward_pass>
self.grad_scaler.scale(loss).backward()
File "/data/users/ezyang/a/pytorch/torch/_tensor.py", line 484, in backward
torch.autograd.backward(
File "/data/users/ezyang/a/pytorch/torch/autograd/__init__.py", line 197, in backward
Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 272, in apply
return user_fn(self, *args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1683, in backward
CompiledFunction.compiled_bw = aot_config.bw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 68, in _wrapped_bw_compiler
return eval_frame.disable(eval_frame.disable(bw_compiler)(*args, **kwargs))
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 380, in bw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 231, in placeholder
sizes, strides = self.static_sizes_strides(example)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 62, in static_sizes_strides
size = [sympy.Integer(i) for i in ex.size()]
AttributeError: 'int' object has no attribute 'size'
While executing %sym_size : [#users=3] = placeholder[target=sym_size]
Original traceback:
Module stack: {'self_cls': "<class 'transformers.models.megatron_bert.modeling_megatron_bert.MegatronBertOnlyMLMHead'>", 'self_cls_predictions': "<class 'transformers.models.megatron_bert.modeling_megatron_bert.MegatronBertLMPredictionHead'>", 'self_cls_predictions_transform': "<class 'transformers.models.megatron_bert.modeling_megatron_bert.MegatronBertPredictionHeadTransform'>", 'self_cls_predictions_transform_dense': "<class 'torch.nn.modules.linear.Linear'>"}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 638, in forward
hidden_states = self.dense(hidden_states)
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 660, in forward
hidden_states = self.transform(hidden_states)
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 672, in forward
prediction_scores = self.predictions(sequence_output)
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 1218, in <graph break in forward>
prediction_scores = self.cls(sequence_output)
Gradient addition node due to multiple use of tensor around:
Module stack: {'self_encoder': "<class 'transformers.models.megatron_bert.modeling_megatron_bert.MegatronBertEncoder'>", 'self_encoder_layer_0': "<class 'transformers.models.megatron_bert.modeling_megatron_bert.MegatronBertLayer'>", 'self_encoder_layer_0_attention': "<class 'transformers.models.megatron_bert.modeling_megatron_bert.MegatronBertAttention'>", 'self_encoder_layer_0_attention_ln': "<class 'torch.nn.modules.normalization.LayerNorm'>"}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 370, in forward
ln_outputs = self.ln(hidden_states)
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 443, in forward
self_attention_outputs = self.attention(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 562, in forward
layer_outputs = layer_module(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 976, in <graph break in forward>
encoder_outputs = self.encoder(
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 4 graph(s) covering 1105 ops
cuda train MegatronBertForQuestionAnswering ERROR:common:'int' object has no attribute 'size'
While executing %sym_size : [#users=1] = placeholder[target=sym_size]
Original traceback:
Module stack: {'self_qa_outputs': "<class 'torch.nn.modules.linear.Linear'>"}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 1813, in <graph break in forward>
logits = self.qa_outputs(sequence_output)
Gradient addition node due to multiple use of tensor around:
Module stack: {'self_encoder': "<class 'transformers.models.megatron_bert.modeling_megatron_bert.MegatronBertEncoder'>", 'self_encoder_layer_0': "<class 'transformers.models.megatron_bert.modeling_megatron_bert.MegatronBertLayer'>", 'self_encoder_layer_0_attention': "<class 'transformers.models.megatron_bert.modeling_megatron_bert.MegatronBertAttention'>", 'self_encoder_layer_0_attention_ln': "<class 'torch.nn.modules.normalization.LayerNorm'>"}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 370, in forward
ln_outputs = self.ln(hidden_states)
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 443, in forward
self_attention_outputs = self.attention(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 562, in forward
layer_outputs = layer_module(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 976, in <graph break in forward>
encoder_outputs = self.encoder(
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 485, in <graph break in forward_and_backward_pass>
self.grad_scaler.scale(loss).backward()
File "/data/users/ezyang/a/pytorch/torch/_tensor.py", line 484, in backward
torch.autograd.backward(
File "/data/users/ezyang/a/pytorch/torch/autograd/__init__.py", line 197, in backward
Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 272, in apply
return user_fn(self, *args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1683, in backward
CompiledFunction.compiled_bw = aot_config.bw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 68, in _wrapped_bw_compiler
return eval_frame.disable(eval_frame.disable(bw_compiler)(*args, **kwargs))
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 380, in bw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 231, in placeholder
sizes, strides = self.static_sizes_strides(example)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 62, in static_sizes_strides
size = [sympy.Integer(i) for i in ex.size()]
AttributeError: 'int' object has no attribute 'size'
While executing %sym_size : [#users=1] = placeholder[target=sym_size]
Original traceback:
Module stack: {'self_qa_outputs': "<class 'torch.nn.modules.linear.Linear'>"}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 1813, in <graph break in forward>
logits = self.qa_outputs(sequence_output)
Gradient addition node due to multiple use of tensor around:
Module stack: {'self_encoder': "<class 'transformers.models.megatron_bert.modeling_megatron_bert.MegatronBertEncoder'>", 'self_encoder_layer_0': "<class 'transformers.models.megatron_bert.modeling_megatron_bert.MegatronBertLayer'>", 'self_encoder_layer_0_attention': "<class 'transformers.models.megatron_bert.modeling_megatron_bert.MegatronBertAttention'>", 'self_encoder_layer_0_attention_ln': "<class 'torch.nn.modules.normalization.LayerNorm'>"}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 370, in forward
ln_outputs = self.ln(hidden_states)
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 443, in forward
self_attention_outputs = self.attention(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 562, in forward
layer_outputs = layer_module(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/megatron_bert/modeling_megatron_bert.py", line 976, in <graph break in forward>
encoder_outputs = self.encoder(
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 4 graph(s) covering 1111 ops
cuda train MobileBertForMaskedLM ERROR:common:compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 952, in aot_dispatch_base
compiled_fw = aot_config.fw_compiler(fw_module, flat_args)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 437, in codegen
self.init_wrapper_code()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 431, in init_wrapper_code
self.wrapper_code = WrapperCodeGen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 302, in __init__
self.write_prefix()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 335, in write_prefix
V.graph.sizevars.codegen(self.wrapper_call, V.graph.graph_inputs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 494, in codegen
assert not needed
AssertionError
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 1089, in forward
outputs = self.mobilebert(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 332, in wrapper
self.output.compile_subgraph(self, reason=reason)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 7 ops
cuda train MobileBertForQuestionAnswering ERROR:common:compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 952, in aot_dispatch_base
compiled_fw = aot_config.fw_compiler(fw_module, flat_args)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 437, in codegen
self.init_wrapper_code()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 431, in init_wrapper_code
self.wrapper_code = WrapperCodeGen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 302, in __init__
self.write_prefix()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 335, in write_prefix
V.graph.sizevars.codegen(self.wrapper_call, V.graph.graph_inputs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 494, in codegen
assert not needed
AssertionError
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/mobilebert/modeling_mobilebert.py", line 1395, in forward
outputs = self.mobilebert(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 332, in wrapper
self.output.compile_subgraph(self, reason=reason)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 7 ops
cuda train OPTForCausalLM [2022-12-12 07:20:19,445] torch._inductor.ir: [WARNING] Using FallbackKernel: aten.cumsum
[2022-12-12 07:20:20,338] torch._inductor.ir: [WARNING] DeviceCopy
ERROR:common:compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 952, in aot_dispatch_base
compiled_fw = aot_config.fw_compiler(fw_module, flat_args)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 437, in codegen
self.init_wrapper_code()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 431, in init_wrapper_code
self.wrapper_code = WrapperCodeGen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 302, in __init__
self.write_prefix()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/wrapper.py", line 335, in write_prefix
V.graph.sizevars.codegen(self.wrapper_call, V.graph.graph_inputs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 494, in codegen
assert not needed
AssertionError
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/opt/modeling_opt.py", line 918, in forward
outputs = self.model.decoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/opt/modeling_opt.py", line 622, in forward
attention_mask = self._prepare_decoder_attention_mask(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 476, in compile_subgraph
self.compile_and_call_fx_graph(tx, list(reversed(stack_values)), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised AssertionError:
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 32 ops
cuda train PLBartForCausalLM [2022-12-12 07:20:34,579] torch._inductor.ir: [WARNING] DeviceCopy
ERROR:common:compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 440, in codegen
self.scheduler.codegen()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/scheduler.py", line 1129, in codegen
self.get_backend(device).codegen_nodes(node.get_nodes())
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1284, in codegen_nodes
return self.codegen_node_schedule(node_schedule, numel, rnumel)
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1336, in codegen_node_schedule
src_code = kernel.codegen_kernel()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1035, in codegen_kernel
"signature": dict(enumerate(map(signature_of, signature))),
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 57, in signature_of
return JITFunction._key_of(V.graph.sizevars.size_hint(arg.expr))
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 367, in size_hint
return int(out)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/plbart/modeling_plbart.py", line 1680, in forward
outputs = self.model.decoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/plbart/modeling_plbart.py", line 1070, in forward
layer_outputs = decoder_layer(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/plbart/modeling_plbart.py", line 424, in forward
hidden_states, self_attn_weights, present_key_value = self.self_attn(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 516, in step
self.output.compile_subgraph(self, partial_convert=True)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 2 graph(s) covering 39 ops
cuda train PLBartForConditionalGeneration ERROR:common:compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 440, in codegen
self.scheduler.codegen()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/scheduler.py", line 1129, in codegen
self.get_backend(device).codegen_nodes(node.get_nodes())
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1284, in codegen_nodes
return self.codegen_node_schedule(node_schedule, numel, rnumel)
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1336, in codegen_node_schedule
src_code = kernel.codegen_kernel()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1035, in codegen_kernel
"signature": dict(enumerate(map(signature_of, signature))),
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 57, in signature_of
return JITFunction._key_of(V.graph.sizevars.size_hint(arg.expr))
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 367, in size_hint
return int(out)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/plbart/modeling_plbart.py", line 1314, in forward
outputs = self.model(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/plbart/modeling_plbart.py", line 1182, in forward
encoder_outputs = self.encoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/plbart/modeling_plbart.py", line 817, in forward
layer_outputs = encoder_layer(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/plbart/modeling_plbart.py", line 328, in forward
hidden_states, attn_weights, _ = self.self_attn(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 516, in step
self.output.compile_subgraph(self, partial_convert=True)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 2 graph(s) covering 39 ops
WARNING:__main__:Sequence Length not defined for PegasusForCausalLM. Choosing 128 arbitrarily
cuda train PegasusForCausalLM ERROR:common:compile_fx raised LoweringException: RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s0
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s0 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
target: aten.arange.default
args[0]: s0
kwargs: {'device': device(type='cpu'), 'pin_memory': False}
While executing %arange : [#users=2] = call_function[target=torch.ops.aten.arange.default](args = (%sym_size,), kwargs = {device: cpu, pin_memory: False})
Original traceback:
Module stack: {}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 84, in _make_causal_mask
mask_cond = torch.arange(mask.size(-1))
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 877, in _prepare_decoder_attention_mask
combined_attention_mask = _make_causal_mask(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1023, in forward
attention_mask = self._prepare_decoder_attention_mask(
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 272, in call_function
out = lowerings[target](*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 221, in wrapped
return decomp_fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 1318, in arange
return fallback_arange(
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 991, in handler
TensorBox.create, ir.FallbackKernel.create(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 3011, in create
) = cls.process_kernel(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 2406, in process_kernel
example_output = kernel(*new_args, **new_kwargs)
File "/data/users/ezyang/a/pytorch/torch/_ops.py", line 500, in __call__
return self._op(*args, **kwargs or {})
File "/data/users/ezyang/a/pytorch/torch/_inductor/overrides.py", line 37, in __torch_function__
return func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_ops.py", line 500, in __call__
return self._op(*args, **kwargs or {})
RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s0
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s0 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 275, in call_function
raise LoweringException(e, target, args, kwargs) from e
torch._inductor.exc.LoweringException: RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s0
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s0 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
target: aten.arange.default
args[0]: s0
kwargs: {'device': device(type='cpu'), 'pin_memory': False}
While executing %arange : [#users=2] = call_function[target=torch.ops.aten.arange.default](args = (%sym_size,), kwargs = {device: cpu, pin_memory: False})
Original traceback:
Module stack: {}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 84, in _make_causal_mask
mask_cond = torch.arange(mask.size(-1))
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 877, in _prepare_decoder_attention_mask
combined_attention_mask = _make_causal_mask(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1023, in forward
attention_mask = self._prepare_decoder_attention_mask(
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1659, in forward
outputs = self.model.decoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 332, in wrapper
self.output.compile_subgraph(self, reason=reason)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised LoweringException: RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s0
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s0 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
target: aten.arange.default
args[0]: s0
kwargs: {'device': device(type='cpu'), 'pin_memory': False}
While executing %arange : [#users=2] = call_function[target=torch.ops.aten.arange.default](args = (%sym_size,), kwargs = {device: cpu, pin_memory: False})
Original traceback:
Module stack: {}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 84, in _make_causal_mask
mask_cond = torch.arange(mask.size(-1))
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 877, in _prepare_decoder_attention_mask
combined_attention_mask = _make_causal_mask(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1023, in forward
attention_mask = self._prepare_decoder_attention_mask(
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 21 ops
WARNING:__main__:Sequence Length not defined for PegasusForConditionalGeneration. Choosing 128 arbitrarily
cuda train PegasusForConditionalGeneration ERROR:common:compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 440, in codegen
self.scheduler.codegen()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/scheduler.py", line 1129, in codegen
self.get_backend(device).codegen_nodes(node.get_nodes())
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1284, in codegen_nodes
return self.codegen_node_schedule(node_schedule, numel, rnumel)
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1336, in codegen_node_schedule
src_code = kernel.codegen_kernel()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1035, in codegen_kernel
"signature": dict(enumerate(map(signature_of, signature))),
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 57, in signature_of
return JITFunction._key_of(V.graph.sizevars.size_hint(arg.expr))
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 367, in size_hint
return int(out)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1399, in forward
outputs = self.model(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 1238, in forward
encoder_outputs = self.encoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 763, in forward
embed_pos = self.embed_positions(input_shape)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 807, in <graph break in forward>
layer_outputs = encoder_layer(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/pegasus/modeling_pegasus.py", line 329, in forward
hidden_states, attn_weights, _ = self.self_attn(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 516, in step
self.output.compile_subgraph(self, partial_convert=True)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 3 graph(s) covering 32 ops
If you want to use `RobertaLMHeadModel` as a standalone, add `is_decoder=True.`
cuda train RobertaForCausalLM [2022-12-12 07:22:10,493] torch._inductor.ir: [WARNING] Using FallbackKernel: aten.cumsum
ERROR:common:compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 440, in codegen
self.scheduler.codegen()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/scheduler.py", line 1129, in codegen
self.get_backend(device).codegen_nodes(node.get_nodes())
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1284, in codegen_nodes
return self.codegen_node_schedule(node_schedule, numel, rnumel)
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1336, in codegen_node_schedule
src_code = kernel.codegen_kernel()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1035, in codegen_kernel
"signature": dict(enumerate(map(signature_of, signature))),
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 57, in signature_of
return JITFunction._key_of(V.graph.sizevars.size_hint(arg.expr))
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 367, in size_hint
return int(out)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/roberta/modeling_roberta.py", line 971, in forward
outputs = self.roberta(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/roberta/modeling_roberta.py", line 848, in forward
encoder_outputs = self.encoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 3 graph(s) covering 565 ops
cuda train RobertaForQuestionAnswering [2022-12-12 07:22:48,672] torch._inductor.ir: [WARNING] Using FallbackKernel: aten.cumsum
ERROR:common:compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 440, in codegen
self.scheduler.codegen()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/scheduler.py", line 1129, in codegen
self.get_backend(device).codegen_nodes(node.get_nodes())
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1284, in codegen_nodes
return self.codegen_node_schedule(node_schedule, numel, rnumel)
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1336, in codegen_node_schedule
src_code = kernel.codegen_kernel()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1035, in codegen_kernel
"signature": dict(enumerate(map(signature_of, signature))),
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 57, in signature_of
return JITFunction._key_of(V.graph.sizevars.size_hint(arg.expr))
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 367, in size_hint
return int(out)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/roberta/modeling_roberta.py", line 1513, in forward
outputs = self.roberta(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/roberta/modeling_roberta.py", line 848, in forward
encoder_outputs = self.encoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 3 graph(s) covering 565 ops
WARNING:__main__:Sequence Length not defined for Speech2Text2ForCausalLM. Choosing 128 arbitrarily
cuda train Speech2Text2ForCausalLM ERROR:common:compile_fx raised LoweringException: RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s0
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s0 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
target: aten.arange.default
args[0]: s0
kwargs: {'device': device(type='cpu'), 'pin_memory': False}
While executing %arange : [#users=2] = call_function[target=torch.ops.aten.arange.default](args = (%sym_size,), kwargs = {device: cpu, pin_memory: False})
Original traceback:
Module stack: {}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/speech_to_text_2/modeling_speech_to_text_2.py", line 53, in _make_causal_mask
mask_cond = torch.arange(mask.size(-1))
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/speech_to_text_2/modeling_speech_to_text_2.py", line 496, in _prepare_decoder_attention_mask
combined_attention_mask = _make_causal_mask(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/speech_to_text_2/modeling_speech_to_text_2.py", line 613, in forward
attention_mask = self._prepare_decoder_attention_mask(
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 272, in call_function
out = lowerings[target](*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 221, in wrapped
return decomp_fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 1318, in arange
return fallback_arange(
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 991, in handler
TensorBox.create, ir.FallbackKernel.create(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 3011, in create
) = cls.process_kernel(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 2406, in process_kernel
example_output = kernel(*new_args, **new_kwargs)
File "/data/users/ezyang/a/pytorch/torch/_ops.py", line 500, in __call__
return self._op(*args, **kwargs or {})
File "/data/users/ezyang/a/pytorch/torch/_inductor/overrides.py", line 37, in __torch_function__
return func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_ops.py", line 500, in __call__
return self._op(*args, **kwargs or {})
RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s0
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s0 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 275, in call_function
raise LoweringException(e, target, args, kwargs) from e
torch._inductor.exc.LoweringException: RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s0
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s0 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
target: aten.arange.default
args[0]: s0
kwargs: {'device': device(type='cpu'), 'pin_memory': False}
While executing %arange : [#users=2] = call_function[target=torch.ops.aten.arange.default](args = (%sym_size,), kwargs = {device: cpu, pin_memory: False})
Original traceback:
Module stack: {}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/speech_to_text_2/modeling_speech_to_text_2.py", line 53, in _make_causal_mask
mask_cond = torch.arange(mask.size(-1))
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/speech_to_text_2/modeling_speech_to_text_2.py", line 496, in _prepare_decoder_attention_mask
combined_attention_mask = _make_causal_mask(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/speech_to_text_2/modeling_speech_to_text_2.py", line 613, in forward
attention_mask = self._prepare_decoder_attention_mask(
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/speech_to_text_2/modeling_speech_to_text_2.py", line 910, in forward
outputs = self.model.decoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 332, in wrapper
self.output.compile_subgraph(self, reason=reason)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised LoweringException: RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s0
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s0 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
target: aten.arange.default
args[0]: s0
kwargs: {'device': device(type='cpu'), 'pin_memory': False}
While executing %arange : [#users=2] = call_function[target=torch.ops.aten.arange.default](args = (%sym_size,), kwargs = {device: cpu, pin_memory: False})
Original traceback:
Module stack: {}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/speech_to_text_2/modeling_speech_to_text_2.py", line 53, in _make_causal_mask
mask_cond = torch.arange(mask.size(-1))
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/speech_to_text_2/modeling_speech_to_text_2.py", line 496, in _prepare_decoder_attention_mask
combined_attention_mask = _make_causal_mask(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/speech_to_text_2/modeling_speech_to_text_2.py", line 613, in forward
attention_mask = self._prepare_decoder_attention_mask(
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 21 ops
cuda train T5ForConditionalGeneration WARNING:common:fp64 golden ref were not generated for T5ForConditionalGeneration. Setting accuracy check to cosine
ERROR:common:compile_fx raised LoweringException: RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s0
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s0 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
target: aten.arange.default
args[0]: s0
kwargs: {'dtype': torch.int64, 'device': device(type='cuda', index=0), 'pin_memory': False}
While executing %arange : [#users=2] = call_function[target=torch.ops.aten.arange.default](args = (%sym_size,), kwargs = {dtype: torch.int64, device: cuda:0, pin_memory: False})
Original traceback:
Module stack: {'self_encoder': "<class 'transformers.models.t5.modeling_t5.T5Stack'>", 'self_encoder_block_0': "<class 'transformers.models.t5.modeling_t5.T5Block'>", 'sub0_0': "<class 'transformers.models.t5.modeling_t5.T5LayerSelfAttention'>", 'self_encoder_block_0_layer_0_SelfAttention': "<class 'transformers.models.t5.modeling_t5.T5Attention'>"}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 423, in compute_bias
context_position = torch.arange(query_length, dtype=torch.long, device=device)[:, None]
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 519, in forward
position_bias = self.compute_bias(real_seq_length, key_length, device=scores.device)
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 570, in forward
attention_output = self.SelfAttention(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 664, in forward
self_attention_outputs = self.layer[0](
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1033, in forward
layer_outputs = layer_module(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1601, in forward
encoder_outputs = self.encoder(
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 272, in call_function
out = lowerings[target](*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 221, in wrapped
return decomp_fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 1318, in arange
return fallback_arange(
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 991, in handler
TensorBox.create, ir.FallbackKernel.create(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 3011, in create
) = cls.process_kernel(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 2406, in process_kernel
example_output = kernel(*new_args, **new_kwargs)
File "/data/users/ezyang/a/pytorch/torch/_ops.py", line 500, in __call__
return self._op(*args, **kwargs or {})
File "/data/users/ezyang/a/pytorch/torch/_inductor/overrides.py", line 37, in __torch_function__
return func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_ops.py", line 500, in __call__
return self._op(*args, **kwargs or {})
RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s0
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s0 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 275, in call_function
raise LoweringException(e, target, args, kwargs) from e
torch._inductor.exc.LoweringException: RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s0
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s0 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
target: aten.arange.default
args[0]: s0
kwargs: {'dtype': torch.int64, 'device': device(type='cuda', index=0), 'pin_memory': False}
While executing %arange : [#users=2] = call_function[target=torch.ops.aten.arange.default](args = (%sym_size,), kwargs = {dtype: torch.int64, device: cuda:0, pin_memory: False})
Original traceback:
Module stack: {'self_encoder': "<class 'transformers.models.t5.modeling_t5.T5Stack'>", 'self_encoder_block_0': "<class 'transformers.models.t5.modeling_t5.T5Block'>", 'sub0_0': "<class 'transformers.models.t5.modeling_t5.T5LayerSelfAttention'>", 'self_encoder_block_0_layer_0_SelfAttention': "<class 'transformers.models.t5.modeling_t5.T5Attention'>"}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 423, in compute_bias
context_position = torch.arange(query_length, dtype=torch.long, device=device)[:, None]
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 519, in forward
position_bias = self.compute_bias(real_seq_length, key_length, device=scores.device)
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 570, in forward
attention_output = self.SelfAttention(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 664, in forward
self_attention_outputs = self.layer[0](
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1033, in forward
layer_outputs = layer_module(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1601, in forward
encoder_outputs = self.encoder(
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised LoweringException: RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s0
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s0 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
target: aten.arange.default
args[0]: s0
kwargs: {'dtype': torch.int64, 'device': device(type='cuda', index=0), 'pin_memory': False}
While executing %arange : [#users=2] = call_function[target=torch.ops.aten.arange.default](args = (%sym_size,), kwargs = {dtype: torch.int64, device: cuda:0, pin_memory: False})
Original traceback:
Module stack: {'self_encoder': "<class 'transformers.models.t5.modeling_t5.T5Stack'>", 'self_encoder_block_0': "<class 'transformers.models.t5.modeling_t5.T5Block'>", 'sub0_0': "<class 'transformers.models.t5.modeling_t5.T5LayerSelfAttention'>", 'self_encoder_block_0_layer_0_SelfAttention': "<class 'transformers.models.t5.modeling_t5.T5Attention'>"}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 423, in compute_bias
context_position = torch.arange(query_length, dtype=torch.long, device=device)[:, None]
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 519, in forward
position_bias = self.compute_bias(real_seq_length, key_length, device=scores.device)
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 570, in forward
attention_output = self.SelfAttention(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 664, in forward
self_attention_outputs = self.layer[0](
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1033, in forward
layer_outputs = layer_module(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1601, in forward
encoder_outputs = self.encoder(
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 885 ops
cuda train T5Small WARNING:common:fp64 golden ref were not generated for T5Small. Setting accuracy check to cosine
ERROR:common:compile_fx raised LoweringException: RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s0
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s0 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
target: aten.arange.default
args[0]: s0
kwargs: {'dtype': torch.int64, 'device': device(type='cuda', index=0), 'pin_memory': False}
While executing %arange : [#users=2] = call_function[target=torch.ops.aten.arange.default](args = (%sym_size,), kwargs = {dtype: torch.int64, device: cuda:0, pin_memory: False})
Original traceback:
Module stack: {'self_encoder': "<class 'transformers.models.t5.modeling_t5.T5Stack'>", 'self_encoder_block_0': "<class 'transformers.models.t5.modeling_t5.T5Block'>", 'sub0_0': "<class 'transformers.models.t5.modeling_t5.T5LayerSelfAttention'>", 'self_encoder_block_0_layer_0_SelfAttention': "<class 'transformers.models.t5.modeling_t5.T5Attention'>"}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 423, in compute_bias
context_position = torch.arange(query_length, dtype=torch.long, device=device)[:, None]
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 519, in forward
position_bias = self.compute_bias(real_seq_length, key_length, device=scores.device)
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 570, in forward
attention_output = self.SelfAttention(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 664, in forward
self_attention_outputs = self.layer[0](
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1033, in forward
layer_outputs = layer_module(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1601, in forward
encoder_outputs = self.encoder(
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 272, in call_function
out = lowerings[target](*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 221, in wrapped
return decomp_fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 1318, in arange
return fallback_arange(
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 991, in handler
TensorBox.create, ir.FallbackKernel.create(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 3011, in create
) = cls.process_kernel(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 2406, in process_kernel
example_output = kernel(*new_args, **new_kwargs)
File "/data/users/ezyang/a/pytorch/torch/_ops.py", line 500, in __call__
return self._op(*args, **kwargs or {})
File "/data/users/ezyang/a/pytorch/torch/_inductor/overrides.py", line 37, in __torch_function__
return func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_ops.py", line 500, in __call__
return self._op(*args, **kwargs or {})
RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s0
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s0 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 275, in call_function
raise LoweringException(e, target, args, kwargs) from e
torch._inductor.exc.LoweringException: RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s0
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s0 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
target: aten.arange.default
args[0]: s0
kwargs: {'dtype': torch.int64, 'device': device(type='cuda', index=0), 'pin_memory': False}
While executing %arange : [#users=2] = call_function[target=torch.ops.aten.arange.default](args = (%sym_size,), kwargs = {dtype: torch.int64, device: cuda:0, pin_memory: False})
Original traceback:
Module stack: {'self_encoder': "<class 'transformers.models.t5.modeling_t5.T5Stack'>", 'self_encoder_block_0': "<class 'transformers.models.t5.modeling_t5.T5Block'>", 'sub0_0': "<class 'transformers.models.t5.modeling_t5.T5LayerSelfAttention'>", 'self_encoder_block_0_layer_0_SelfAttention': "<class 'transformers.models.t5.modeling_t5.T5Attention'>"}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 423, in compute_bias
context_position = torch.arange(query_length, dtype=torch.long, device=device)[:, None]
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 519, in forward
position_bias = self.compute_bias(real_seq_length, key_length, device=scores.device)
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 570, in forward
attention_output = self.SelfAttention(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 664, in forward
self_attention_outputs = self.layer[0](
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1033, in forward
layer_outputs = layer_module(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1601, in forward
encoder_outputs = self.encoder(
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised LoweringException: RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s0
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s0 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
target: aten.arange.default
args[0]: s0
kwargs: {'dtype': torch.int64, 'device': device(type='cuda', index=0), 'pin_memory': False}
While executing %arange : [#users=2] = call_function[target=torch.ops.aten.arange.default](args = (%sym_size,), kwargs = {dtype: torch.int64, device: cuda:0, pin_memory: False})
Original traceback:
Module stack: {'self_encoder': "<class 'transformers.models.t5.modeling_t5.T5Stack'>", 'self_encoder_block_0': "<class 'transformers.models.t5.modeling_t5.T5Block'>", 'sub0_0': "<class 'transformers.models.t5.modeling_t5.T5LayerSelfAttention'>", 'self_encoder_block_0_layer_0_SelfAttention': "<class 'transformers.models.t5.modeling_t5.T5Attention'>"}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 423, in compute_bias
context_position = torch.arange(query_length, dtype=torch.long, device=device)[:, None]
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 519, in forward
position_bias = self.compute_bias(real_seq_length, key_length, device=scores.device)
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 570, in forward
attention_output = self.SelfAttention(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 664, in forward
self_attention_outputs = self.layer[0](
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1033, in forward
layer_outputs = layer_module(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1601, in forward
encoder_outputs = self.encoder(
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 885 ops
cuda train TrOCRForCausalLM [2022-12-12 07:25:07,098] torch._inductor.ir: [WARNING] DeviceCopy
ERROR:common:compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 136, in compile_fx_inner
compiled_fn = graph.compile_to_fn()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 503, in compile_to_fn
return self.compile_to_module().call
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 488, in compile_to_module
code = self.codegen()
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 440, in codegen
self.scheduler.codegen()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/scheduler.py", line 1129, in codegen
self.get_backend(device).codegen_nodes(node.get_nodes())
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1284, in codegen_nodes
return self.codegen_node_schedule(node_schedule, numel, rnumel)
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1336, in codegen_node_schedule
src_code = kernel.codegen_kernel()
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 1035, in codegen_kernel
"signature": dict(enumerate(map(signature_of, signature))),
File "/data/users/ezyang/a/pytorch/torch/_inductor/codegen/triton.py", line 57, in signature_of
return JITFunction._key_of(V.graph.sizevars.size_hint(arg.expr))
File "/data/users/ezyang/a/pytorch/torch/_inductor/sizevars.py", line 367, in size_hint
return int(out)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/sympy-1.11.1-py3.9.egg/sympy/core/expr.py", line 320, in __int__
raise TypeError("Cannot convert symbols to int")
TypeError: Cannot convert symbols to int
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/trocr/modeling_trocr.py", line 953, in forward
outputs = self.model.decoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/trocr/modeling_trocr.py", line 715, in forward
layer_outputs = decoder_layer(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/trocr/modeling_trocr.py", line 381, in forward
hidden_states, self_attn_weights, present_key_value = self.self_attn(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 516, in step
self.output.compile_subgraph(self, partial_convert=True)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised TypeError: Cannot convert symbols to int
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 2 graph(s) covering 40 ops
WARNING:__main__:Sequence Length not defined for XGLMForCausalLM. Choosing 128 arbitrarily
cuda train XGLMForCausalLM ERROR:common:compile_fx raised LoweringException: RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s0
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s0 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
target: aten.arange.default
args[0]: s0
kwargs: {'device': device(type='cpu'), 'pin_memory': False}
While executing %arange : [#users=2] = call_function[target=torch.ops.aten.arange.default](args = (%sym_size,), kwargs = {device: cpu, pin_memory: False})
Original traceback:
Module stack: {}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/xglm/modeling_xglm.py", line 124, in _make_causal_mask
mask_cond = torch.arange(mask.size(-1))
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/xglm/modeling_xglm.py", line 578, in _prepare_decoder_attention_mask
combined_attention_mask = _make_causal_mask(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/xglm/modeling_xglm.py", line 701, in forward
attention_mask = self._prepare_decoder_attention_mask(
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 272, in call_function
out = lowerings[target](*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 221, in wrapped
return decomp_fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 1318, in arange
return fallback_arange(
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 991, in handler
TensorBox.create, ir.FallbackKernel.create(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 3011, in create
) = cls.process_kernel(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 2406, in process_kernel
example_output = kernel(*new_args, **new_kwargs)
File "/data/users/ezyang/a/pytorch/torch/_ops.py", line 500, in __call__
return self._op(*args, **kwargs or {})
File "/data/users/ezyang/a/pytorch/torch/_inductor/overrides.py", line 37, in __torch_function__
return func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_ops.py", line 500, in __call__
return self._op(*args, **kwargs or {})
RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s0
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s0 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 275, in call_function
raise LoweringException(e, target, args, kwargs) from e
torch._inductor.exc.LoweringException: RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s0
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s0 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
target: aten.arange.default
args[0]: s0
kwargs: {'device': device(type='cpu'), 'pin_memory': False}
While executing %arange : [#users=2] = call_function[target=torch.ops.aten.arange.default](args = (%sym_size,), kwargs = {device: cpu, pin_memory: False})
Original traceback:
Module stack: {}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/xglm/modeling_xglm.py", line 124, in _make_causal_mask
mask_cond = torch.arange(mask.size(-1))
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/xglm/modeling_xglm.py", line 578, in _prepare_decoder_attention_mask
combined_attention_mask = _make_causal_mask(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/xglm/modeling_xglm.py", line 701, in forward
attention_mask = self._prepare_decoder_attention_mask(
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/xglm/modeling_xglm.py", line 889, in forward
outputs = self.model(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 332, in wrapper
self.output.compile_subgraph(self, reason=reason)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised LoweringException: RuntimeError: Overloaded torch operator invoked from Python failed to many any schema:
aten::arange() expected at most 5 argument(s) but received 7 argument(s). Declaration: aten::arange(Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() expected at most 6 argument(s) but received 7 argument(s). Declaration: aten::arange.start(Scalar start, Scalar end, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
aten::arange() Expected a value of type 'number' for argument 'end' but instead found type 'Symbol'.
Position: 1
Value: s0
Declaration: aten::arange.start_step(Scalar start, Scalar end, Scalar step=1, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
Cast error details: Cannot cast s0 to number
aten::arange() expected at most 4 argument(s) but received 7 argument(s). Declaration: aten::arange.start_out(Scalar start, Scalar end, Scalar step=1, *, Tensor(a!) out) -> Tensor(a!)
aten::arange() expected at most 2 argument(s) but received 7 argument(s). Declaration: aten::arange.out(Scalar end, *, Tensor(a!) out) -> Tensor(a!)
target: aten.arange.default
args[0]: s0
kwargs: {'device': device(type='cpu'), 'pin_memory': False}
While executing %arange : [#users=2] = call_function[target=torch.ops.aten.arange.default](args = (%sym_size,), kwargs = {device: cpu, pin_memory: False})
Original traceback:
Module stack: {}
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/xglm/modeling_xglm.py", line 124, in _make_causal_mask
mask_cond = torch.arange(mask.size(-1))
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/xglm/modeling_xglm.py", line 578, in _prepare_decoder_attention_mask
combined_attention_mask = _make_causal_mask(
| File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/xglm/modeling_xglm.py", line 701, in forward
attention_mask = self._prepare_decoder_attention_mask(
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 21 ops
cuda train XLNetLMHeadModel [2022-12-12 07:27:26,608] torch._inductor.ir: [WARNING] DeviceCopy
[2022-12-12 07:27:26,631] torch._inductor.graph: [WARNING] Creating implicit fallback for:
target: <built-in function sub>
args[0]: 1024
args[1]: 1
ERROR:common:compile_fx raised LoweringException: TypeError: sub expected 2 arguments, got 0
target: <built-in function sub>
args[0]: 1024
args[1]: 1
While executing %sub : [#users=1] = call_function[target=operator.sub](args = (%sym_size_9, 1), kwargs = {})
Original traceback:
None
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 272, in call_function
out = lowerings[target](*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 221, in wrapped
return decomp_fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 991, in handler
TensorBox.create, ir.FallbackKernel.create(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 3011, in create
) = cls.process_kernel(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 2406, in process_kernel
example_output = kernel(*new_args, **new_kwargs)
TypeError: sub expected 2 arguments, got 0
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 275, in call_function
raise LoweringException(e, target, args, kwargs) from e
torch._inductor.exc.LoweringException: TypeError: sub expected 2 arguments, got 0
target: <built-in function sub>
args[0]: 1024
args[1]: 1
While executing %sub : [#users=1] = call_function[target=operator.sub](args = (%sym_size_9, 1), kwargs = {})
Original traceback:
None
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1448, in forward
transformer_outputs = self.transformer(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1207, in forward
pos_emb = self.relative_positional_encoding(qlen, klen, bsz=bsz)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised LoweringException: TypeError: sub expected 2 arguments, got 0
target: <built-in function sub>
args[0]: 1024
args[1]: 1
While executing %sub : [#users=1] = call_function[target=operator.sub](args = (%sym_size_9, 1), kwargs = {})
Original traceback:
None
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 2 graph(s) covering 1009 ops
cuda train YituTechConvBert [2022-12-12 07:28:17,885] torch._inductor.graph: [WARNING] Creating implicit fallback for:
target: <built-in function sub>
args[0]: 520
args[1]: 8
ERROR:common:compile_fx raised LoweringException: TypeError: sub expected 2 arguments, got 0
target: <built-in function sub>
args[0]: 520
args[1]: 8
While executing %sub_1 : [#users=2] = call_function[target=operator.sub](args = (%add_2, 8), kwargs = {})
Original traceback:
None
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 272, in call_function
out = lowerings[target](*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 221, in wrapped
return decomp_fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 991, in handler
TensorBox.create, ir.FallbackKernel.create(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 3011, in create
) = cls.process_kernel(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 2406, in process_kernel
example_output = kernel(*new_args, **new_kwargs)
TypeError: sub expected 2 arguments, got 0
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 275, in call_function
raise LoweringException(e, target, args, kwargs) from e
torch._inductor.exc.LoweringException: TypeError: sub expected 2 arguments, got 0
target: <built-in function sub>
args[0]: 520
args[1]: 8
While executing %sub_1 : [#users=2] = call_function[target=operator.sub](args = (%add_2, 8), kwargs = {})
Original traceback:
None
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 480, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 481, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/huggingface.py", line 483, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/convbert/modeling_convbert.py", line 928, in forward
generator_hidden_states = self.convbert(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/transformers/models/convbert/modeling_convbert.py", line 853, in forward
hidden_states = self.encoder(
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 500, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised LoweringException: TypeError: sub expected 2 arguments, got 0
target: <built-in function sub>
args[0]: 520
args[1]: 8
While executing %sub_1 : [#users=2] = call_function[target=operator.sub](args = (%add_2, 8), kwargs = {})
Original traceback:
None
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 3 graph(s) covering 827 ops
cuda train adv_inception_v3 [2022-12-12 07:29:42,686] torch._inductor.graph: [WARNING] Creating implicit fallback for:
target: <built-in function truediv>
args[0]: 5683456
args[1]: 32
ERROR:common:compile_fx raised LoweringException: TypeError: truediv expected 2 arguments, got 0
target: <built-in function truediv>
args[0]: 5683456
args[1]: 32
While executing %truediv : [#users=2] = call_function[target=operator.truediv](args = (%sym_numel, 32), kwargs = {})
Original traceback:
None
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 272, in call_function
out = lowerings[target](*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 221, in wrapped
return decomp_fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 991, in handler
TensorBox.create, ir.FallbackKernel.create(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 3011, in create
) = cls.process_kernel(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 2406, in process_kernel
example_output = kernel(*new_args, **new_kwargs)
TypeError: truediv expected 2 arguments, got 0
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 275, in call_function
raise LoweringException(e, target, args, kwargs) from e
torch._inductor.exc.LoweringException: TypeError: truediv expected 2 arguments, got 0
target: <built-in function truediv>
args[0]: 5683456
args[1]: 32
While executing %truediv : [#users=2] = call_function[target=operator.truediv](args = (%sym_numel, 32), kwargs = {})
Original traceback:
None
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/timm_models.py", line 312, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/timm_models.py", line 313, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/timm_models.py", line 315, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 332, in catch_errors
return callback(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 479, in _convert_frame
result = inner_convert(frame, cache_size, hooks)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 103, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 339, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 398, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/convert_frame.py", line 385, in transform
tracer.run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1676, in run
super().run()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 528, in run
and self.step()
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 496, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/symbolic_convert.py", line 1738, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 476, in compile_subgraph
self.compile_and_call_fx_graph(tx, list(reversed(stack_values)), root)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 547, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 628, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fx raised LoweringException: TypeError: truediv expected 2 arguments, got 0
target: <built-in function truediv>
args[0]: 5683456
args[1]: 32
While executing %truediv : [#users=2] = call_function[target=operator.truediv](args = (%sym_numel, 32), kwargs = {})
Original traceback:
None
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 0 graph(s) covering 313 ops
cuda train beit_base_patch16_224 ERROR:common:name 's0' is not defined
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/timm_models.py", line 312, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/timm_models.py", line 313, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/timm_models.py", line 315, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/timm/models/beit.py", line 341, in forward
def forward(self, x):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2340, in forward
return compiled_fn(full_args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 887, in g
return f(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1903, in debug_compiled_function
return compiled_function(*args)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1718, in compiled_function
all_outs = CompiledFunction.apply(*args_with_synthetic_bases)
File "/data/users/ezyang/a/pytorch/torch/autograd/function.py", line 419, in apply
return super().apply(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1581, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 912, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 199, in run
return model(new_inputs)
File "/tmp/torchinductor_ezyang/mw/cmwozk3udaucc75xv6rddb425ftttk25fh6vyeqgzrylvedhek2u.py", line 1424, in call
return (buf360, primals_2, primals_7, primals_8, primals_13, primals_14, primals_19, primals_20, primals_25, primals_26, primals_31, primals_32, primals_37, primals_38, primals_43, primals_44, primals_49, primals_50, primals_55, primals_56, primals_61, primals_62, primals_67, primals_68, primals_73, primals_74, primals_76, primals_77, primals_80, primals_81, primals_86, primals_87, primals_90, primals_91, primals_96, primals_97, primals_100, primals_101, primals_106, primals_107, primals_110, primals_111, primals_116, primals_117, primals_120, primals_121, primals_126, primals_127, primals_130, primals_131, primals_136, primals_137, primals_140, primals_141, primals_146, primals_147, primals_150, primals_151, primals_156, primals_157, primals_160, primals_161, primals_166, primals_167, primals_170, primals_171, primals_176, primals_177, primals_180, primals_181, primals_186, primals_187, primals_190, primals_191, primals_196, primals_197, primals_224, buf3, buf7, buf8, as_strided(buf15, (96, 197, 64), (12608, 64, 1)), as_strided(primals_201, (38809, ), (1, )), buf20, as_strided(buf23, (1576, 768), (768, 1)), buf24, buf28, as_strided(buf30, (8, 197, 3072), (605184, 3072, 1)), buf32, buf37, as_strided(buf44, (96, 197, 64), (12608, 64, 1)), as_strided(primals_203, (38809, ), (1, )), buf49, as_strided(buf52, (1576, 768), (768, 1)), buf53, buf57, as_strided(buf59, (8, 197, 3072), (605184, 3072, 1)), buf61, buf66, as_strided(buf73, (96, 197, 64), (12608, 64, 1)), as_strided(primals_205, (38809, ), (1, )), buf78, as_strided(buf81, (1576, 768), (768, 1)), buf82, buf86, as_strided(buf88, (8, 197, 3072), (605184, 3072, 1)), buf90, buf95, as_strided(buf102, (96, 197, 64), (12608, 64, 1)), as_strided(primals_207, (38809, ), (1, )), buf107, as_strided(buf110, (1576, 768), (768, 1)), buf111, buf115, as_strided(buf117, (8, 197, 3072), (605184, 3072, 1)), buf119, buf124, as_strided(buf131, (96, 197, 64), (12608, 64, 1)), as_strided(primals_209, (38809, ), (1, )), buf136, as_strided(buf139, (1576, 768), (768, 1)), buf140, buf144, as_strided(buf146, (8, 197, 3072), (605184, 3072, 1)), buf148, buf153, as_strided(buf160, (96, 197, 64), (12608, 64, 1)), as_strided(primals_211, (38809, ), (1, )), buf165, as_strided(buf168, (1576, 768), (768, 1)), buf169, buf173, as_strided(buf175, (8, 197, 3072), (605184, 3072, 1)), buf177, buf182, as_strided(buf189, (96, 197, 64), (12608, 64, 1)), as_strided(primals_213, (38809, ), (1, )), buf194, as_strided(buf197, (1576, 768), (768, 1)), buf198, buf202, as_strided(buf204, (8, 197, 3072), (605184, 3072, 1)), buf206, buf211, as_strided(buf218, (96, 197, 64), (12608, 64, 1)), as_strided(primals_215, (38809, ), (1, )), buf223, as_strided(buf226, (1576, 768), (768, 1)), buf227, buf231, as_strided(buf233, (8, 197, 3072), (605184, 3072, 1)), buf235, buf240, as_strided(buf247, (96, 197, 64), (12608, 64, 1)), as_strided(primals_217, (38809, ), (1, )), buf252, as_strided(buf255, (1576, 768), (768, 1)), buf256, buf260, as_strided(buf262, (8, 197, 3072), (605184, 3072, 1)), buf264, buf269, as_strided(buf276, (96, 197, 64), (12608, 64, 1)), as_strided(primals_219, (38809, ), (1, )), buf281, as_strided(buf284, (1576, 768), (768, 1)), buf285, buf289, as_strided(buf291, (8, 197, 3072), (605184, 3072, 1)), buf293, buf298, as_strided(buf305, (96, 197, 64), (12608, 64, 1)), as_strided(primals_221, (38809, ), (1, )), buf310, as_strided(buf313, (1576, 768), (768, 1)), buf314, buf318, as_strided(buf320, (8, 197, 3072), (605184, 3072, 1)), buf322, buf327, as_strided(buf334, (96, 197, 64), (12608, 64, 1)), as_strided(primals_223, (38809, ), (1, )), buf339, as_strided(buf342, (1576, 768), (768, 1)), buf343, buf347, as_strided(buf349, (8, 197, 3072), (605184, 3072, 1)), buf351, buf358, as_strided(primals_198, (1000, 768), (768, 1)), buf361, as_strided(primals_194, (768, 3072), (3072, 1)), as_strided(primals_192, (3072, 768), (768, 1)), buf362, as_strided(primals_188, (768, 768), (768, 1)), as_strided(buf340, (96, 64, 197), (12608, 1, 64)), as_strided(buf335, (96, 197, 64), (12608, 1, 197)), as_strided(primals_71, (2304, 768), (768, 1)), buf363, as_strided(primals_184, (768, 3072), (3072, 1)), as_strided(primals_182, (3072, 768), (768, 1)), buf364, as_strided(primals_178, (768, 768), (768, 1)), as_strided(buf311, (96, 64, 197), (12608, 1, 64)), as_strided(buf306, (96, 197, 64), (12608, 1, 197)), as_strided(primals_65, (2304, 768), (768, 1)), buf365, as_strided(primals_174, (768, 3072), (3072, 1)), as_strided(primals_172, (3072, 768), (768, 1)), buf366, as_strided(primals_168, (768, 768), (768, 1)), as_strided(buf282, (96, 64, 197), (12608, 1, 64)), as_strided(buf277, (96, 197, 64), (12608, 1, 197)), as_strided(primals_59, (2304, 768), (768, 1)), buf367, as_strided(primals_164, (768, 3072), (3072, 1)), as_strided(primals_162, (3072, 768), (768, 1)), buf368, as_strided(primals_158, (768, 768), (768, 1)), as_strided(buf253, (96, 64, 197), (12608, 1, 64)), as_strided(buf248, (96, 197, 64), (12608, 1, 197)), as_strided(primals_53, (2304, 768), (768, 1)), buf369, as_strided(primals_154, (768, 3072), (3072, 1)), as_strided(primals_152, (3072, 768), (768, 1)), buf370, as_strided(primals_148, (768, 768), (768, 1)), as_strided(buf224, (96, 64, 197), (12608, 1, 64)), as_strided(buf219, (96, 197, 64), (12608, 1, 197)), as_strided(primals_47, (2304, 768), (768, 1)), buf371, as_strided(primals_144, (768, 3072), (3072, 1)), as_strided(primals_142, (3072, 768), (768, 1)), buf372, as_strided(primals_138, (768, 768), (768, 1)), as_strided(buf195, (96, 64, 197), (12608, 1, 64)), as_strided(buf190, (96, 197, 64), (12608, 1, 197)), as_strided(primals_41, (2304, 768), (768, 1)), buf373, as_strided(primals_134, (768, 3072), (3072, 1)), as_strided(primals_132, (3072, 768), (768, 1)), buf374, as_strided(primals_128, (768, 768), (768, 1)), as_strided(buf166, (96, 64, 197), (12608, 1, 64)), as_strided(buf161, (96, 197, 64), (12608, 1, 197)), as_strided(primals_35, (2304, 768), (768, 1)), buf375, as_strided(primals_124, (768, 3072), (3072, 1)), as_strided(primals_122, (3072, 768), (768, 1)), buf376, as_strided(primals_118, (768, 768), (768, 1)), as_strided(buf137, (96, 64, 197), (12608, 1, 64)), as_strided(buf132, (96, 197, 64), (12608, 1, 197)), as_strided(primals_29, (2304, 768), (768, 1)), buf377, as_strided(primals_114, (768, 3072), (3072, 1)), as_strided(primals_112, (3072, 768), (768, 1)), buf378, as_strided(primals_108, (768, 768), (768, 1)), as_strided(buf108, (96, 64, 197), (12608, 1, 64)), as_strided(buf103, (96, 197, 64), (12608, 1, 197)), as_strided(primals_23, (2304, 768), (768, 1)), buf379, as_strided(primals_104, (768, 3072), (3072, 1)), as_strided(primals_102, (3072, 768), (768, 1)), buf380, as_strided(primals_98, (768, 768), (768, 1)), as_strided(buf79, (96, 64, 197), (12608, 1, 64)), as_strided(buf74, (96, 197, 64), (12608, 1, 197)), as_strided(primals_17, (2304, 768), (768, 1)), buf381, as_strided(primals_94, (768, 3072), (3072, 1)), as_strided(primals_92, (3072, 768), (768, 1)), buf382, as_strided(primals_88, (768, 768), (768, 1)), as_strided(buf50, (96, 64, 197), (12608, 1, 64)), as_strided(buf45, (96, 197, 64), (12608, 1, 197)), as_strided(primals_11, (2304, 768), (768, 1)), buf383, as_strided(primals_84, (768, 3072), (3072, 1)), as_strided(primals_82, (3072, 768), (768, 1)), buf384, as_strided(primals_78, (768, 768), (768, 1)), as_strided(buf21, (96, 64, 197), (12608, 1, 64)), as_strided(buf16, (96, 197, 64), (12608, 1, 197)), as_strided(primals_5, (2304, 768), (768, 1)), s0, 14, 14, 197, 197*s0, 197, 12, 12*s0, 12, 768, 197, 197*s0, 197, 12, 12*s0, 12, 768, 197, 197*s0, 197, 12, 12*s0, 12, 768, 197, 197*s0, 197, 12, 12*s0, 12, 768, 197, 197*s0, 197, 12, 12*s0, 12, 768, 197, 197*s0, 197, 12, 12*s0, 12, 768, 197, 197*s0, 197, 12, 12*s0, 12, 768, 197, 197*s0, 197, 12, 12*s0, 12, 768, 197, 197*s0, 197, 12, 12*s0, 12, 768, 197, 197*s0, 197, 12, 12*s0, 12, 768, 197, 197*s0, 197, 12, 12*s0, 12, 768, 197, 197*s0, 197, 12, 12*s0, 12, 768, 197, 197*s0, s0, 196, 196, )
NameError: name 's0' is not defined
TorchDynamo optimized model failed to run because of following error
FAIL
Dynamo produced 1 graph(s) covering 513 ops
cuda train botnet26t_256 [2022-12-12 07:30:48,793] torch._inductor.graph: [WARNING] Creating implicit fallback for:
target: <built-in function truediv>
args[0]: 3145728
args[1]: 24
ERROR:common:compile_fx raised LoweringException: TypeError: truediv expected 2 arguments, got 0
target: <built-in function truediv>
args[0]: 3145728
args[1]: 24
While executing %truediv : [#users=2] = call_function[target=operator.truediv](args = (%sym_numel, 24), kwargs = {})
Original traceback:
None
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 272, in call_function
out = lowerings[target](*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 221, in wrapped
return decomp_fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/lowering.py", line 991, in handler
TensorBox.create, ir.FallbackKernel.create(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 3011, in create
) = cls.process_kernel(kernel, *args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/ir.py", line 2406, in process_kernel
example_output = kernel(*new_args, **new_kwargs)
TypeError: truediv expected 2 arguments, got 0
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/torch/_dynamo/output_graph.py", line 623, in call_user_compiler
compiled_fn = compiler_fn(gm, self.fake_example_inputs())
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 917, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 394, in compile_fx
return aot_autograd(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/optimizations/training.py", line 78, in compiler_fn
cg = aot_module_simplified(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2326, in aot_module_simplified
compiled_fn = create_aot_dispatcher_function(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 2023, in create_aot_dispatcher_function
compiled_fn = compiler_fn(flat_fn, fake_flat_tensor_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1293, in aot_wrapper_dedupe
return compiler_fn(flat_fn, leaf_flat_args, aot_config)
File "/data/users/ezyang/a/pytorch/torch/_functorch/aot_autograd.py", line 1540, in aot_dispatch_autograd
compiled_fw_func = aot_config.fw_compiler(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 369, in fw_compiler
return inner_compile(
File "/data/users/ezyang/a/pytorch/torch/_dynamo/debug_utils.py", line 494, in debug_wrapper
compiled_fn = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/debug.py", line 224, in inner
return fn(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/contextlib.py", line 79, in inner
return func(*args, **kwds)
File "/data/users/ezyang/a/pytorch/torch/_inductor/compile_fx.py", line 135, in compile_fx_inner
graph.run(*example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 147, in run
return super().run(*args)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 345, in run_node
result = super().run_node(n)
File "/data/users/ezyang/a/pytorch/torch/fx/interpreter.py", line 171, in run_node
return getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/a/pytorch/torch/_inductor/graph.py", line 275, in call_function
raise LoweringException(e, target, args, kwargs) from e
torch._inductor.exc.LoweringException: TypeError: truediv expected 2 arguments, got 0
target: <built-in function truediv>
args[0]: 3145728
args[1]: 24
While executing %truediv : [#users=2] = call_function[target=operator.truediv](args = (%sym_numel, 24), kwargs = {})
Original traceback:
None
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(model_copy, example_inputs)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/eval_frame.py", line 211, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/common.py", line 1040, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/timm_models.py", line 312, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/timm_models.py", line 313, in <graph break in forward_and_backward_pass>
self.optimizer_zero_grad(mod)
File "/data/users/ezyang/a/pytorch/benchmarks/dynamo/timm_models.py", line 315, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/a/pytorch/torch/nn/modules/module.py", line 1482, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ezyang/local/a/pytorch-env/lib/python3.9/site-packages/timm/models/byobnet.py", line 1559, in forward
x = self.forward_features(x)
File "/data/users/ezyang/a/pytorch/torch/_dynamo/
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment