This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| cuda train BERT_pytorch torchdynamo.symbolic_convert: [WARNING] Graph break: call_function in skip_files /scratch/whc/work/torchdynamo/torchdynamo/utils.py from user code at File "benchmarks/torchbench.py", line 341, in forward_and_backward_pass | |
| cloned_inputs = clone_inputs(inputs) | |
| torchdynamo.symbolic_convert: [WARNING] Graph break: call_method NNModuleVariable() zero_grad [ConstantVariable(bool)] {} from user code at File "benchmarks/torchbench.py", line 342, in <graph break in forward_and_backward_pass> | |
| mod.zero_grad(True) | |
| torchdynamo.symbolic_convert: [WARNING] Graph break: Dynamic slicing not supported from user code at File "/scratch/whc/work/torchbenchmark/torchbenchmark/models/BERT_pytorch/bert_pytorch/model/bert.py", line 43, in forward | |
| x = self.embedding(x, segment_info) | |
| File "/scratch/whc/work/torchbenchmark/torchbenchmark/models/BERT_pytorch/bert_pytorch/model/embedding/bert.py", line 32, in forward | |
| x = self.token(sequence) + self.position(sequence) |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| cuda train BERT_pytorch torchdynamo.symbolic_convert: [WARNING] Graph break: call_function in skip_files /scratch/whc/work/torchdynamo/torchdynamo/utils.py from user code at File "benchmarks/torchbench.py", line 341, in forward_and_backward_pass | |
| cloned_inputs = clone_inputs(inputs) | |
| torchdynamo.symbolic_convert: [WARNING] Graph break: call_method NNModuleVariable() zero_grad [ConstantVariable(bool)] {} from user code at File "benchmarks/torchbench.py", line 342, in <graph break in forward_and_backward_pass> | |
| mod.zero_grad(True) | |
| torchdynamo.symbolic_convert: [WARNING] Graph break: Dynamic slicing not supported from user code at File "/scratch/whc/work/torchbenchmark/torchbenchmark/models/BERT_pytorch/bert_pytorch/model/bert.py", line 43, in forward | |
| x = self.embedding(x, segment_info) | |
| File "/scratch/whc/work/torchbenchmark/torchbenchmark/models/BERT_pytorch/bert_pytorch/model/embedding/bert.py", line 32, in forward | |
| x = self.token(sequence) + self.position(sequence) |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| cuda train BERT_pytorch torchdynamo.symbolic_convert: [WARNING] Graph break: call_function in skip_files /scratch/whc/work/torchdynamo/torchdynamo/utils.py from user code at File "benchmarks/torchbench.py", line 341, in forward_and_backward_pass | |
| cloned_inputs = clone_inputs(inputs) | |
| torchdynamo.symbolic_convert: [WARNING] Graph break: call_method NNModuleVariable() zero_grad [ConstantVariable(bool)] {} from user code at File "benchmarks/torchbench.py", line 342, in <graph break in forward_and_backward_pass> | |
| mod.zero_grad(True) | |
| torchdynamo.symbolic_convert: [WARNING] Graph break: Dynamic slicing not supported from user code at File "/scratch/whc/work/torchbenchmark/torchbenchmark/models/BERT_pytorch/bert_pytorch/model/bert.py", line 43, in forward | |
| x = self.embedding(x, segment_info) | |
| File "/scratch/whc/work/torchbenchmark/torchbenchmark/models/BERT_pytorch/bert_pytorch/model/embedding/bert.py", line 32, in forward | |
| x = self.token(sequence) + self.position(sequence) |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| cuda train BERT_pytorch torchdynamo.symbolic_convert: [WARNING] Graph break: call_function in skip_files /scratch/whc/work/torchdynamo/torchdynamo/utils.py from user code at File "benchmarks/torchbench.py", line 341, in forward_and_backward_pass | |
| cloned_inputs = clone_inputs(inputs) | |
| torchdynamo.symbolic_convert: [WARNING] Graph break: call_method NNModuleVariable() zero_grad [ConstantVariable(bool)] {} from user code at File "benchmarks/torchbench.py", line 342, in <graph break in forward_and_backward_pass> | |
| mod.zero_grad(True) | |
| torchdynamo.symbolic_convert: [WARNING] Graph break: Dynamic slicing not supported from user code at File "/scratch/whc/work/torchbenchmark/torchbenchmark/models/BERT_pytorch/bert_pytorch/model/bert.py", line 43, in forward | |
| x = self.embedding(x, segment_info) | |
| File "/scratch/whc/work/torchbenchmark/torchbenchmark/models/BERT_pytorch/bert_pytorch/model/embedding/bert.py", line 32, in forward | |
| x = self.token(sequence) + self.position(sequence) |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| cuda train BERT_pytorch torchdynamo.symbolic_convert: [WARNING] Graph break: call_function in skip_files /scratch/whc/work/torchdynamo/torchdynamo/utils.py from user code at File "benchmarks/torchbench.py", line 341, in forward_and_backward_pass | |
| cloned_inputs = clone_inputs(inputs) | |
| torchdynamo.symbolic_convert: [WARNING] Graph break: call_method NNModuleVariable() zero_grad [ConstantVariable(bool)] {} from user code at File "benchmarks/torchbench.py", line 342, in <graph break in forward_and_backward_pass> | |
| mod.zero_grad(True) | |
| torchdynamo.symbolic_convert: [WARNING] Graph break: Dynamic slicing not supported from user code at File "/scratch/whc/work/torchbenchmark/torchbenchmark/models/BERT_pytorch/bert_pytorch/model/bert.py", line 43, in forward | |
| x = self.embedding(x, segment_info) | |
| File "/scratch/whc/work/torchbenchmark/torchbenchmark/models/BERT_pytorch/bert_pytorch/model/embedding/bert.py", line 32, in forward | |
| x = self.token(sequence) + self.position(sequence) |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| cuda train BERT_pytorch torchdynamo.symbolic_convert: [WARNING] Graph break: call_function in skip_files /scratch/whc/work/torchdynamo/torchdynamo/utils.py from user code at File "benchmarks/torchbench.py", line 341, in forward_and_backward_pass | |
| cloned_inputs = clone_inputs(inputs) | |
| torchdynamo.symbolic_convert: [WARNING] Graph break: call_method NNModuleVariable() zero_grad [ConstantVariable(bool)] {} from user code at File "benchmarks/torchbench.py", line 342, in <graph break in forward_and_backward_pass> | |
| mod.zero_grad(True) | |
| torchdynamo.symbolic_convert: [WARNING] Graph break: Dynamic slicing not supported from user code at File "/scratch/whc/work/torchbenchmark/torchbenchmark/models/BERT_pytorch/bert_pytorch/model/bert.py", line 43, in forward | |
| x = self.embedding(x, segment_info) | |
| File "/scratch/whc/work/torchbenchmark/torchbenchmark/models/BERT_pytorch/bert_pytorch/model/embedding/bert.py", line 32, in forward | |
| x = self.token(sequence) + self.position(sequence) |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| cuda train BERT_pytorch torchdynamo.symbolic_convert: [WARNING] Graph break: call_function in skip_files /scratch/whc/work/torchdynamo/torchdynamo/utils.py from user code at File "benchmarks/torchbench.py", line 341, in forward_and_backward_pass | |
| cloned_inputs = clone_inputs(inputs) | |
| torchdynamo.symbolic_convert: [WARNING] Graph break: call_method NNModuleVariable() zero_grad [ConstantVariable(bool)] {} from user code at File "benchmarks/torchbench.py", line 342, in <graph break in forward_and_backward_pass> | |
| mod.zero_grad(True) | |
| torchdynamo.symbolic_convert: [WARNING] Graph break: Dynamic slicing not supported from user code at File "/scratch/whc/work/torchbenchmark/torchbenchmark/models/BERT_pytorch/bert_pytorch/model/bert.py", line 43, in forward | |
| x = self.embedding(x, segment_info) | |
| File "/scratch/whc/work/torchbenchmark/torchbenchmark/models/BERT_pytorch/bert_pytorch/model/embedding/bert.py", line 32, in forward | |
| x = self.token(sequence) + self.position(sequence) |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| cuda train BERT_pytorch torchdynamo.symbolic_convert: [WARNING] Graph break: call_function in skip_files /scratch/whc/work/torchdynamo/torchdynamo/utils.py from user code at File "benchmarks/torchbench.py", line 341, in forward_and_backward_pass | |
| cloned_inputs = clone_inputs(inputs) | |
| torchdynamo.symbolic_convert: [WARNING] Graph break: call_method NNModuleVariable() zero_grad [ConstantVariable(bool)] {} from user code at File "benchmarks/torchbench.py", line 342, in <graph break in forward_and_backward_pass> | |
| mod.zero_grad(True) | |
| torchdynamo.symbolic_convert: [WARNING] Graph break: Dynamic slicing not supported from user code at File "/scratch/whc/work/torchbenchmark/torchbenchmark/models/BERT_pytorch/bert_pytorch/model/bert.py", line 43, in forward | |
| x = self.embedding(x, segment_info) | |
| File "/scratch/whc/work/torchbenchmark/torchbenchmark/models/BERT_pytorch/bert_pytorch/model/embedding/bert.py", line 32, in forward | |
| x = self.token(sequence) + self.position(sequence) |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| cuda train BERT_pytorch torchdynamo.symbolic_convert: [WARNING] Graph break: call_function in skip_files /scratch/whc/work/torchdynamo/torchdynamo/utils.py from user code at File "benchmarks/torchbench.py", line 341, in forward_and_backward_pass | |
| cloned_inputs = clone_inputs(inputs) | |
| torchdynamo.symbolic_convert: [WARNING] Graph break: call_method NNModuleVariable() zero_grad [ConstantVariable(bool)] {} from user code at File "benchmarks/torchbench.py", line 342, in <graph break in forward_and_backward_pass> | |
| mod.zero_grad(True) | |
| torchdynamo.symbolic_convert: [WARNING] Graph break: Dynamic slicing not supported from user code at File "/scratch/whc/work/torchbenchmark/torchbenchmark/models/BERT_pytorch/bert_pytorch/model/bert.py", line 43, in forward | |
| x = self.embedding(x, segment_info) | |
| File "/scratch/whc/work/torchbenchmark/torchbenchmark/models/BERT_pytorch/bert_pytorch/model/embedding/bert.py", line 32, in forward | |
| x = self.token(sequence) + self.position(sequence) |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| cuda train BERT_pytorch torchdynamo.symbolic_convert: [WARNING] Graph break: call_function in skip_files /scratch/whc/work/torchdynamo/torchdynamo/utils.py from user code at File "benchmarks/torchbench.py", line 341, in forward_and_backward_pass | |
| cloned_inputs = clone_inputs(inputs) | |
| torchdynamo.symbolic_convert: [WARNING] Graph break: call_method NNModuleVariable() zero_grad [ConstantVariable(bool)] {} from user code at File "benchmarks/torchbench.py", line 342, in <graph break in forward_and_backward_pass> | |
| mod.zero_grad(True) | |
| torchdynamo.symbolic_convert: [WARNING] Graph break: Dynamic slicing not supported from user code at File "/scratch/whc/work/torchbenchmark/torchbenchmark/models/BERT_pytorch/bert_pytorch/model/bert.py", line 43, in forward | |
| x = self.embedding(x, segment_info) | |
| File "/scratch/whc/work/torchbenchmark/torchbenchmark/models/BERT_pytorch/bert_pytorch/model/embedding/bert.py", line 32, in forward | |
| x = self.token(sequence) + self.position(sequence) |