Created
July 6, 2023 04:27
-
-
Save fxmarty/5b5dbfac7ef4a3ea746effdf146584b6 to your computer and use it in GitHub Desktop.
log_dynamo_inductor_aot
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Some weights of the model checkpoint at fxmarty/tiny-llama-fast-tokenizer were not used when initializing LlamaForCausalLM: ['model.layers.1.self_attn.q_proj.weight', 'model.layers.0.self_attn.v_proj.weight', 'model.layers.1.self_attn.v_proj.weight', 'model.layers.0.self_attn.k_proj.weight', 'model.layers.1.self_attn.k_proj.weight', 'model.layers.0.self_attn.q_proj.weight'] | |
- This IS expected if you are initializing LlamaForCausalLM from the checkpoint of a model trained on another task or with another architecture (e.g. initializing a BertForSequenceClassification model from a BertForPreTraining model). | |
- This IS NOT expected if you are initializing LlamaForCausalLM from the checkpoint of a model that you expect to be exactly identical (initializing a BertForSequenceClassification model from a BertForSequenceClassification model). | |
Some weights of LlamaForCausalLM were not initialized from the model checkpoint at fxmarty/tiny-llama-fast-tokenizer and are newly initialized: ['model.layers.0.self_attn.qkv_proj.weight', 'model.layers.1.self_attn.qkv_proj.weight'] | |
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference. | |
0%| | 0/1 [00:00<?, ?it/s] | |
0%| | 0/1 [00:00<?, ?it/s][A | |
0%| | 0/1 [00:00<?, ?it/s][A[A[2023-07-06 04:20:38,056] torch._dynamo.eval_frame: [DEBUG] skipping __init__ /home/ubuntu/miniconda3/lib/python3.10/contextlib.py | |
[2023-07-06 04:20:38,056] torch._dynamo.eval_frame: [DEBUG] skipping __enter__ /home/ubuntu/miniconda3/lib/python3.10/contextlib.py | |
[2023-07-06 04:20:38,056] torch._dynamo.eval_frame: [DEBUG] skipping __init__ /home/ubuntu/miniconda3/lib/python3.10/contextlib.py | |
[2023-07-06 04:20:38,056] torch._dynamo.eval_frame: [DEBUG] skipping __enter__ /home/ubuntu/miniconda3/lib/python3.10/contextlib.py | |
[2023-07-06 04:20:38,056] torch._dynamo.eval_frame: [DEBUG] skipping enable_dynamic /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py | |
[2023-07-06 04:20:38,057] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo start tracing forward /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:644 | |
[2023-07-06 04:20:38,139] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['input_ids'] (1, 1000) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None] | |
[2023-07-06 04:20:38,140] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['attention_mask'] (1, 1003) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None] | |
[2023-07-06 04:20:38,141] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['position_ids'] (1, 1000) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None] | |
[2023-07-06 04:20:38,142] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['past_key_values'][0] (2, 1, 4, 1003, 4) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None, None, None] | |
[2023-07-06 04:20:38,142] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['past_key_values'][1] (2, 1, 4, 1003, 4) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None, None, None] | |
[2023-07-06 04:20:38,143] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['valid_past_index'] () [] [] | |
[2023-07-06 04:20:38,144] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:685 | |
[2023-07-06 04:20:38,144] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST output_attentions [] | |
[2023-07-06 04:20:38,144] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:38,144] torch._dynamo.symbolic_convert: [DEBUG] TRACE IS_OP 1 [ConstantVariable(NoneType), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:38,144] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 12 [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,144] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,144] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR config [NNModuleVariable()] | |
[2023-07-06 04:20:38,145] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR output_attentions [HFPretrainedConfigVariable()] | |
[2023-07-06 04:20:38,145] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST output_attentions [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,145] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:687 | |
[2023-07-06 04:20:38,145] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST output_hidden_states [] | |
[2023-07-06 04:20:38,145] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:38,145] torch._dynamo.symbolic_convert: [DEBUG] TRACE IS_OP 1 [ConstantVariable(NoneType), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:38,146] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 32 [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,146] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,146] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR config [NNModuleVariable()] | |
[2023-07-06 04:20:38,146] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR output_hidden_states [HFPretrainedConfigVariable()] | |
[2023-07-06 04:20:38,146] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:686 | |
[2023-07-06 04:20:38,146] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST output_hidden_states [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,146] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:689 | |
[2023-07-06 04:20:38,146] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST return_dict [] | |
[2023-07-06 04:20:38,146] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,146] torch._dynamo.symbolic_convert: [DEBUG] TRACE IS_OP 1 [ConstantVariable(bool), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:38,147] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 52 [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,147] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST return_dict [] | |
[2023-07-06 04:20:38,147] torch._dynamo.symbolic_convert: [DEBUG] TRACE JUMP_FORWARD 58 [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,147] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST return_dict [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,147] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:692 | |
[2023-07-06 04:20:38,147] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,147] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR model [NNModuleVariable()] | |
[2023-07-06 04:20:38,147] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:693 | |
[2023-07-06 04:20:38,147] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST input_ids [NNModuleVariable()] | |
[2023-07-06 04:20:38,147] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:694 | |
[2023-07-06 04:20:38,147] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST attention_mask [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:38,148] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:695 | |
[2023-07-06 04:20:38,148] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST position_ids [NNModuleVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:38,148] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:696 | |
[2023-07-06 04:20:38,148] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST past_key_values [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:38,148] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:697 | |
[2023-07-06 04:20:38,148] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST inputs_embeds [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ListVariable()] | |
[2023-07-06 04:20:38,148] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:698 | |
[2023-07-06 04:20:38,148] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST output_attentions [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ListVariable(), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:38,148] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:699 | |
[2023-07-06 04:20:38,148] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST output_hidden_states [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ListVariable(), ConstantVariable(NoneType), ConstantVariable(bool)] | |
[2023-07-06 04:20:38,148] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:700 | |
[2023-07-06 04:20:38,148] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST return_dict [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ListVariable(), ConstantVariable(NoneType), ConstantVariable(bool), ConstantVariable(bool)] | |
[2023-07-06 04:20:38,148] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:701 | |
[2023-07-06 04:20:38,148] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ListVariable(), ConstantVariable(NoneType), ConstantVariable(bool), ConstantVariable(bool), ConstantVariable(bool)] | |
[2023-07-06 04:20:38,148] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:692 | |
[2023-07-06 04:20:38,148] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('input_ids', 'attention_mask', 'position_ids', 'past_key_values', 'inputs_embeds', 'output_attentions', 'output_hidden_states', 'return_dict', 'valid_past_index') [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ListVariable(), ConstantVariable(NoneType), ConstantVariable(bool), ConstantVariable(bool), ConstantVariable(bool), TensorVariable()] | |
[2023-07-06 04:20:38,148] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 9 [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ListVariable(), ConstantVariable(NoneType), ConstantVariable(bool), ConstantVariable(bool), ConstantVariable(bool), TensorVariable(), ConstantVariable(tuple)] | |
[2023-07-06 04:20:38,149] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:38,151] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:38,151] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:38,153] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:38,156] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:38,156] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:38,156] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:38,156] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,156] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:38,157] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function LlamaModel.forward at 0x7fb8d0c46cb0>, NNModuleVariable())] | |
[2023-07-06 04:20:38,157] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:38,157] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,157] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:38,227] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,227] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,227] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:38,228] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,228] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,228] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:38,228] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,228] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,228] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:38,228] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,228] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:38,228] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:38,229] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:38,229] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,229] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:38,229] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:38,229] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:38,229] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,229] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:38,229] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:38,229] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:38,229] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,229] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:38,229] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:38,229] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:38,229] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,229] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:38,229] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:38,229] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function LlamaModel.forward at 0x7fb8d0c46cb0>, NNModuleVariable())] | |
[2023-07-06 04:20:38,229] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function LlamaModel.forward at 0x7fb8d0c46cb0>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:38,230] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function LlamaModel.forward at 0x7fb8d0c46cb0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:38,230] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function LlamaModel.forward at 0x7fb8d0c46cb0>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:38,230] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function LlamaModel.forward at 0x7fb8d0c46cb0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:38,231] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:38,231] torch._dynamo.symbolic_convert: [DEBUG] FAILED INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:38,231] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:38,231] torch._dynamo.symbolic_convert: [DEBUG] break_graph_if_unsupported triggered compile | |
Traceback (most recent call last): | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2187, in inline_call_ | |
sub_locals, closure_cells = func.bind_args(parent, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 187, in bind_args | |
[ | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 188, in <listcomp> | |
wrap(val=arg, source=source) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 62, in wrap_bound_arg | |
return VariableBuilder(tx, source=source, **options)(val) | |
TypeError: VariableBuilder.__init__() got an unexpected keyword argument 'guards' | |
During handling of the above exception, another exception occurred: | |
Traceback (most recent call last): | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 390, in wrapper | |
return inner_fn(self, inst) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 1152, in CALL_FUNCTION_KW | |
self.call_function(fn, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 559, in call_function | |
self.push(fn.call_function(self, args, kwargs)) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/nn_module.py", line 333, in call_function | |
return tx.inline_user_function_return( | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 595, in inline_user_function_return | |
result = InliningInstructionTranslator.inline_call(self, fn, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2132, in inline_call | |
return cls.inline_call_(parent, func, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2229, in inline_call_ | |
tracer.run() | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 708, in run | |
and self.step() | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 668, in step | |
getattr(self, inst.opname)(inst) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 390, in wrapper | |
return inner_fn(self, inst) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 1140, in CALL_FUNCTION_EX | |
self.call_function(fn, argsvars.items, kwargsvars.items) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 559, in call_function | |
self.push(fn.call_function(self, args, kwargs)) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 333, in call_function | |
return super().call_function(tx, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 287, in call_function | |
return super().call_function(tx, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 120, in call_function | |
return tx.inline_user_function_return( | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 595, in inline_user_function_return | |
result = InliningInstructionTranslator.inline_call(self, fn, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2132, in inline_call | |
return cls.inline_call_(parent, func, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2190, in inline_call_ | |
raise ArgsMismatchError( | |
torch._dynamo.exc.ArgsMismatchError: VariableBuilder.__init__() got an unexpected keyword argument 'guards'. | |
func = 'forward' /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:477, args = [<class 'trfs_fast.llama.LlamaModel'>], kwargs = OrderedDict([('input_ids', TensorVariable()), ('attention_mask', TensorVariable()), ('position_ids', TensorVariable()), ('past_key_values', ListVariable()), ('inputs_embeds', ConstantVariable(NoneType)), ('output_attentions', ConstantVariable(bool)), ('output_hidden_states', ConstantVariable(bool)), ('return_dict', ConstantVariable(bool)), ('valid_past_index', TensorVariable())]) | |
[2023-07-06 04:20:38,235] torch._dynamo.symbolic_convert.__graph_breaks: [DEBUG] Graph break: VariableBuilder.__init__() got an unexpected keyword argument 'guards'. | |
func = 'forward' /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:477, args = [<class 'trfs_fast.llama.LlamaModel'>], kwargs = OrderedDict([('input_ids', TensorVariable()), ('attention_mask', TensorVariable()), ('position_ids', TensorVariable()), ('past_key_values', ListVariable()), ('inputs_embeds', ConstantVariable(NoneType)), ('output_attentions', ConstantVariable(bool)), ('output_hidden_states', ConstantVariable(bool)), ('return_dict', ConstantVariable(bool)), ('valid_past_index', TensorVariable())]) from user code at File "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 692, in forward | |
outputs = self.model( | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1514, in _call_impl | |
return forward_call(*args, **kwargs) | |
[2023-07-06 04:20:38,235] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:38,235] torch._dynamo.output_graph: [DEBUG] COMPILING GRAPH due to GraphCompileReason(reason="VariableBuilder.__init__() got an unexpected keyword argument 'guards'.\n func = 'forward' /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:477, args = [<class 'trfs_fast.llama.LlamaModel'>], kwargs = OrderedDict([('input_ids', TensorVariable()), ('attention_mask', TensorVariable()), ('position_ids', TensorVariable()), ('past_key_values', ListVariable()), ('inputs_embeds', ConstantVariable(NoneType)), ('output_attentions', ConstantVariable(bool)), ('output_hidden_states', ConstantVariable(bool)), ('return_dict', ConstantVariable(bool)), ('valid_past_index', TensorVariable())])", user_stack=[<FrameSummary file /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py, line 692 in forward>, <FrameSummary file /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py, line 1514 in _call_impl>], graph_break=True) | |
[2023-07-06 04:20:38,295] torch._dynamo.convert_frame.__guards: [DEBUG] GUARDS: | |
___check_obj_id(L['self'], 140431662761824) | |
hasattr(L['input_ids'], '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['return_dict'], 7698144) | |
hasattr(L['position_ids'], '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['inputs_embeds'], 7649088) | |
hasattr(L['attention_mask'], '_dynamo_dynamic_indices') == False | |
___check_type_id(L['past_key_values'], 7662688) | |
len(L['past_key_values']) == 2 | |
hasattr(L['valid_past_index'], '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['output_attentions'], 7649088) | |
hasattr(L['past_key_values'][0], '_dynamo_dynamic_indices') == False | |
hasattr(L['past_key_values'][1], '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['output_hidden_states'], 7649088) | |
not ___is_grad_enabled() | |
not ___are_deterministic_algorithms_enabled() | |
utils_device.CURRENT_DEVICE == None | |
[2023-07-06 04:20:38,295] torch._dynamo.eval_frame: [DEBUG] skipping __getattr__ /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py | |
[2023-07-06 04:20:38,295] torch._dynamo.convert_frame: [DEBUG] skipping because no torch.* __getattribute__ /home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/configuration_utils.py 258 | |
[2023-07-06 04:20:38,295] torch._dynamo.eval_frame: [DEBUG] skipping _wrapped_call_impl /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py | |
[2023-07-06 04:20:38,295] torch._dynamo.eval_frame: [DEBUG] skipping _call_impl /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py | |
[2023-07-06 04:20:38,296] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo start tracing forward /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:477 | |
[2023-07-06 04:20:38,297] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['input_ids'] (1, 1000) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None] | |
[2023-07-06 04:20:38,298] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['attention_mask'] (1, 1003) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None] | |
[2023-07-06 04:20:38,299] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['position_ids'] (1, 1000) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None] | |
[2023-07-06 04:20:38,299] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['past_key_values'][0] (2, 1, 4, 1003, 4) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None, None, None] | |
[2023-07-06 04:20:38,300] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['past_key_values'][1] (2, 1, 4, 1003, 4) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None, None, None] | |
[2023-07-06 04:20:38,301] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['valid_past_index'] () [] [] | |
[2023-07-06 04:20:38,301] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:490 | |
[2023-07-06 04:20:38,301] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_DEREF output_attentions [] | |
[2023-07-06 04:20:38,302] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,302] torch._dynamo.symbolic_convert: [DEBUG] TRACE IS_OP 1 [ConstantVariable(bool), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:38,302] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 12 [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,302] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_DEREF output_attentions [] | |
[2023-07-06 04:20:38,302] torch._dynamo.symbolic_convert: [DEBUG] TRACE JUMP_FORWARD 18 [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,302] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_DEREF output_attentions [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,302] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:492 | |
[2023-07-06 04:20:38,302] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST output_hidden_states [] | |
[2023-07-06 04:20:38,302] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,302] torch._dynamo.symbolic_convert: [DEBUG] TRACE IS_OP 1 [ConstantVariable(bool), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:38,302] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 32 [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,302] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST output_hidden_states [] | |
[2023-07-06 04:20:38,302] torch._dynamo.symbolic_convert: [DEBUG] TRACE JUMP_FORWARD 38 [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,302] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:491 | |
[2023-07-06 04:20:38,302] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST output_hidden_states [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,302] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:494 | |
[2023-07-06 04:20:38,302] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST return_dict [] | |
[2023-07-06 04:20:38,302] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,302] torch._dynamo.symbolic_convert: [DEBUG] TRACE IS_OP 1 [ConstantVariable(bool), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:38,303] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 52 [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,303] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST return_dict [] | |
[2023-07-06 04:20:38,303] torch._dynamo.symbolic_convert: [DEBUG] TRACE JUMP_FORWARD 58 [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,303] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST return_dict [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,303] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:497 | |
[2023-07-06 04:20:38,303] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST input_ids [] | |
[2023-07-06 04:20:38,303] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [TensorVariable()] | |
[2023-07-06 04:20:38,303] torch._dynamo.symbolic_convert: [DEBUG] TRACE IS_OP 1 [TensorVariable(), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:38,303] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 84 [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,303] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST inputs_embeds [] | |
[2023-07-06 04:20:38,303] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:38,303] torch._dynamo.symbolic_convert: [DEBUG] TRACE IS_OP 1 [ConstantVariable(NoneType), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:38,303] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 84 [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,303] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:499 | |
[2023-07-06 04:20:38,303] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST input_ids [] | |
[2023-07-06 04:20:38,303] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [TensorVariable()] | |
[2023-07-06 04:20:38,303] torch._dynamo.symbolic_convert: [DEBUG] TRACE IS_OP 1 [TensorVariable(), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:38,303] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 104 [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,303] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:500 | |
[2023-07-06 04:20:38,304] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST input_ids [] | |
[2023-07-06 04:20:38,304] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR shape [TensorVariable()] | |
[2023-07-06 04:20:38,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE UNPACK_SEQUENCE 2 [ShapeVariable()] | |
[2023-07-06 04:20:38,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST batch_size [ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:38,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST seq_length [ConstantVariable(int)] | |
[2023-07-06 04:20:38,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE JUMP_FORWARD 134 [] | |
[2023-07-06 04:20:38,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:506 | |
[2023-07-06 04:20:38,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST seq_length [] | |
[2023-07-06 04:20:38,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST seq_length_with_past [ConstantVariable(int)] | |
[2023-07-06 04:20:38,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:507 | |
[2023-07-06 04:20:38,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [] | |
[2023-07-06 04:20:38,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST past_key_values_length [TensorVariable()] | |
[2023-07-06 04:20:38,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:509 | |
[2023-07-06 04:20:38,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST position_ids [] | |
[2023-07-06 04:20:38,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [TensorVariable()] | |
[2023-07-06 04:20:38,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE IS_OP 0 [TensorVariable(), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:38,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 214 [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,319] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:516 | |
[2023-07-06 04:20:38,319] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST position_ids [] | |
[2023-07-06 04:20:38,319] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR view [TensorVariable()] | |
[2023-07-06 04:20:38,319] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [GetAttrVariable(TensorVariable(), view)] | |
[2023-07-06 04:20:38,319] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST seq_length [GetAttrVariable(TensorVariable(), view), ConstantVariable(int)] | |
[2023-07-06 04:20:38,319] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 2 [GetAttrVariable(TensorVariable(), view), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:38,322] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR long [TensorVariable()] | |
[2023-07-06 04:20:38,322] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [GetAttrVariable(TensorVariable(), long)] | |
[2023-07-06 04:20:38,322] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST position_ids [TensorVariable()] | |
[2023-07-06 04:20:38,322] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:518 | |
[2023-07-06 04:20:38,322] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST inputs_embeds [] | |
[2023-07-06 04:20:38,322] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:38,323] torch._dynamo.symbolic_convert: [DEBUG] TRACE IS_OP 0 [ConstantVariable(NoneType), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:38,323] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 248 [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,323] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:519 | |
[2023-07-06 04:20:38,323] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,323] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR embed_tokens [NNModuleVariable()] | |
[2023-07-06 04:20:38,323] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST input_ids [NNModuleVariable()] | |
[2023-07-06 04:20:38,323] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:38,333] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST inputs_embeds [TensorVariable()] | |
[2023-07-06 04:20:38,333] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:521 | |
[2023-07-06 04:20:38,333] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST attention_mask [] | |
[2023-07-06 04:20:38,333] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [TensorVariable()] | |
[2023-07-06 04:20:38,333] torch._dynamo.symbolic_convert: [DEBUG] TRACE IS_OP 0 [TensorVariable(), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:38,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 280 [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:527 | |
[2023-07-06 04:20:38,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST batch_size [] | |
[2023-07-06 04:20:38,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [ConstantVariable(int)] | |
[2023-07-06 04:20:38,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE COMPARE_OP > [ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:38,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 308 [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:532 | |
[2023-07-06 04:20:38,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST inputs_embeds [] | |
[2023-07-06 04:20:38,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:38,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:535 | |
[2023-07-06 04:20:38,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST output_hidden_states [] | |
[2023-07-06 04:20:38,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 320 [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [] | |
[2023-07-06 04:20:38,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST all_hidden_states [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:38,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:536 | |
[2023-07-06 04:20:38,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_DEREF output_attentions [] | |
[2023-07-06 04:20:38,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 332 [ConstantVariable(bool)] | |
[2023-07-06 04:20:38,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [] | |
[2023-07-06 04:20:38,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST all_self_attns [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:38,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:537 | |
[2023-07-06 04:20:38,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST () [] | |
[2023-07-06 04:20:38,335] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST next_decoder_cache [TupleVariable()] | |
[2023-07-06 04:20:38,335] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:539 | |
[2023-07-06 04:20:38,335] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL enumerate [] | |
[2023-07-06 04:20:38,335] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [BuiltinVariable(enumerate)] | |
[2023-07-06 04:20:38,335] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR layers [BuiltinVariable(enumerate), NNModuleVariable()] | |
[2023-07-06 04:20:38,335] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [BuiltinVariable(enumerate), NNModuleVariable()] | |
[2023-07-06 04:20:38,336] torch._dynamo.symbolic_convert: [DEBUG] TRACE GET_ITER None [TupleVariable()] | |
[2023-07-06 04:20:38,336] torch._dynamo.symbolic_convert: [DEBUG] TRACE FOR_ITER 516 [ListIteratorVariable()] | |
[2023-07-06 04:20:38,336] torch._dynamo.symbolic_convert: [DEBUG] TRACE UNPACK_SEQUENCE 2 [ListIteratorVariable(), TupleVariable()] | |
[2023-07-06 04:20:38,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST idx [ListIteratorVariable(), NNModuleVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:38,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST decoder_layer [ListIteratorVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:38,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:540 | |
[2023-07-06 04:20:38,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST output_hidden_states [ListIteratorVariable()] | |
[2023-07-06 04:20:38,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 372 [ListIteratorVariable(), ConstantVariable(bool)] | |
[2023-07-06 04:20:38,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:543 | |
[2023-07-06 04:20:38,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST past_key_values [ListIteratorVariable()] | |
[2023-07-06 04:20:38,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [ListIteratorVariable(), ListVariable()] | |
[2023-07-06 04:20:38,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE IS_OP 1 [ListIteratorVariable(), ListVariable(), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:38,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 388 [ListIteratorVariable(), ConstantVariable(bool)] | |
[2023-07-06 04:20:38,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST past_key_values [ListIteratorVariable()] | |
[2023-07-06 04:20:38,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST idx [ListIteratorVariable(), ListVariable()] | |
[2023-07-06 04:20:38,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [ListIteratorVariable(), ListVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:38,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE JUMP_FORWARD 390 [ListIteratorVariable(), TensorVariable()] | |
[2023-07-06 04:20:38,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST past_key_value [ListIteratorVariable(), TensorVariable()] | |
[2023-07-06 04:20:38,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:545 | |
[2023-07-06 04:20:38,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [ListIteratorVariable()] | |
[2023-07-06 04:20:38,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR gradient_checkpointing [ListIteratorVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:38,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 444 [ListIteratorVariable(), ConstantVariable(bool)] | |
[2023-07-06 04:20:38,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:562 | |
[2023-07-06 04:20:38,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST decoder_layer [ListIteratorVariable()] | |
[2023-07-06 04:20:38,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:563 | |
[2023-07-06 04:20:38,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [ListIteratorVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:38,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:564 | |
[2023-07-06 04:20:38,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST attention_mask [ListIteratorVariable(), NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:38,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:565 | |
[2023-07-06 04:20:38,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST position_ids [ListIteratorVariable(), NNModuleVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:38,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:566 | |
[2023-07-06 04:20:38,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST past_key_value [ListIteratorVariable(), NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:38,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:567 | |
[2023-07-06 04:20:38,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_DEREF output_attentions [ListIteratorVariable(), NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:38,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:568 | |
[2023-07-06 04:20:38,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [ListIteratorVariable(), NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ConstantVariable(bool)] | |
[2023-07-06 04:20:38,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:562 | |
[2023-07-06 04:20:38,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('attention_mask', 'position_ids', 'past_key_value', 'output_attentions', 'valid_past_index') [ListIteratorVariable(), NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ConstantVariable(bool), TensorVariable()] | |
[2023-07-06 04:20:38,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 6 [ListIteratorVariable(), NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ConstantVariable(bool), TensorVariable(), ConstantVariable(tuple)] | |
[2023-07-06 04:20:38,339] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:38,340] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:38,340] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:38,340] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:38,340] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:38,341] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:38,341] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:38,341] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,341] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:38,341] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function LlamaDecoderLayer.forward at 0x7fb8d0c467a0>, NNModuleVariable())] | |
[2023-07-06 04:20:38,341] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:38,341] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,341] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:38,342] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,342] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,342] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:38,342] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,342] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,342] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:38,342] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,342] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,342] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:38,343] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,343] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:38,343] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:38,343] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:38,343] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,343] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:38,343] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:38,343] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:38,343] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,343] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:38,343] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:38,343] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:38,343] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,343] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:38,343] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:38,344] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:38,344] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,344] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:38,344] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:38,344] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function LlamaDecoderLayer.forward at 0x7fb8d0c467a0>, NNModuleVariable())] | |
[2023-07-06 04:20:38,344] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function LlamaDecoderLayer.forward at 0x7fb8d0c467a0>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:38,344] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function LlamaDecoderLayer.forward at 0x7fb8d0c467a0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:38,344] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function LlamaDecoderLayer.forward at 0x7fb8d0c467a0>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:38,344] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function LlamaDecoderLayer.forward at 0x7fb8d0c467a0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:38,345] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:38,345] torch._dynamo.symbolic_convert: [DEBUG] FAILED INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:38,345] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:38,345] torch._dynamo.symbolic_convert: [INFO] Skipping frame because there is a graph break in a for/while loop | |
<FrameSummary file /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py, line 562 in forward> | |
[2023-07-06 04:20:38,345] torch._dynamo.convert_frame: [DEBUG] Skipping frame Skipping frame because there is a graph break in a for/while loop | |
<FrameSummary file /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py, line 562 in forward> forward /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py 477 | |
[2023-07-06 04:20:38,345] torch._dynamo.eval_frame: [DEBUG] skipping forward /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/sparse.py | |
[2023-07-06 04:20:38,345] torch._dynamo.eval_frame: [DEBUG] skipping embedding /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/functional.py | |
[2023-07-06 04:20:38,352] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo start tracing __iter__ /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/container.py:315 | |
[2023-07-06 04:20:38,352] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/container.py:317 | |
[2023-07-06 04:20:38,353] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL iter [] | |
[2023-07-06 04:20:38,354] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [BuiltinVariable(iter)] | |
[2023-07-06 04:20:38,354] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _modules [BuiltinVariable(iter), NNModuleVariable()] | |
[2023-07-06 04:20:38,354] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR values [BuiltinVariable(iter), ConstDictVariable()] | |
[2023-07-06 04:20:38,355] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [BuiltinVariable(iter), GetAttrVariable(ConstDictVariable(), values)] | |
[2023-07-06 04:20:38,355] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [BuiltinVariable(iter), TupleVariable()] | |
[2023-07-06 04:20:38,355] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [ListIteratorVariable()] | |
[2023-07-06 04:20:38,355] torch._dynamo.convert_frame: [DEBUG] Skipping frame because no content in function call __iter__ /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/container.py 315 | |
[2023-07-06 04:20:38,355] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo start tracing forward /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:261 | |
[2023-07-06 04:20:38,356] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['hidden_states'] (1, 1000, 16) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None] | |
[2023-07-06 04:20:38,357] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['attention_mask'] (1, 1003) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None] | |
[2023-07-06 04:20:38,357] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['position_ids'] (1, 1000) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None] | |
[2023-07-06 04:20:38,359] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['past_key_value'] (2, 1, 4, 1003, 4) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None, None, None] | |
[2023-07-06 04:20:38,360] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['valid_past_index'] () [] [] | |
[2023-07-06 04:20:38,360] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:281 | |
[2023-07-06 04:20:38,360] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:38,360] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST residual [TensorVariable()] | |
[2023-07-06 04:20:38,360] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:283 | |
[2023-07-06 04:20:38,360] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,360] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR input_layernorm [NNModuleVariable()] | |
[2023-07-06 04:20:38,361] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [NNModuleVariable()] | |
[2023-07-06 04:20:38,361] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:38,361] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:38,362] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:38,363] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:38,363] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:38,363] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:38,363] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:38,363] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:38,363] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,363] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:38,364] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable())] | |
[2023-07-06 04:20:38,364] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:38,364] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,364] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:38,364] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,364] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,364] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:38,364] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,364] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,364] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:38,365] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,365] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,365] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:38,365] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,365] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:38,365] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:38,365] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:38,365] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,365] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:38,365] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:38,366] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:38,366] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,366] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:38,366] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:38,366] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:38,366] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,366] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:38,366] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:38,366] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:38,366] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,366] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:38,366] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:38,366] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable())] | |
[2023-07-06 04:20:38,367] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:38,367] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:38,367] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:38,367] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:38,367] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object forward at 0x7fb8cddd7ec0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 83> | |
[2023-07-06 04:20:38,367] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:84 | |
[2023-07-06 04:20:38,367] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:38,367] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR dtype [TensorVariable()] | |
[2023-07-06 04:20:38,370] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST input_dtype [TorchVariable(torch.float16)] | |
[2023-07-06 04:20:38,370] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85 | |
[2023-07-06 04:20:38,370] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:38,370] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable()] | |
[2023-07-06 04:20:38,370] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:38,370] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR float32 [GetAttrVariable(TensorVariable(), to), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:38,371] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), to), ConstantVariable(dtype)] | |
[2023-07-06 04:20:38,372] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR pow [TensorVariable()] | |
[2023-07-06 04:20:38,372] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [GetAttrVariable(TensorVariable(), pow)] | |
[2023-07-06 04:20:38,372] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), pow), ConstantVariable(int)] | |
[2023-07-06 04:20:38,374] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR mean [TensorVariable()] | |
[2023-07-06 04:20:38,374] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [GetAttrVariable(TensorVariable(), mean)] | |
[2023-07-06 04:20:38,374] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST True [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int)] | |
[2023-07-06 04:20:38,374] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('keepdim',) [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int), ConstantVariable(bool)] | |
[2023-07-06 04:20:38,374] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 2 [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int), ConstantVariable(bool), ConstantVariable(tuple)] | |
[2023-07-06 04:20:38,376] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST variance [TensorVariable()] | |
[2023-07-06 04:20:38,376] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86 | |
[2023-07-06 04:20:38,376] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:38,376] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [TensorVariable()] | |
[2023-07-06 04:20:38,376] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR rsqrt [TensorVariable(), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:38,376] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST variance [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:38,377] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:38,377] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR variance_epsilon [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:38,377] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(float)] | |
[2023-07-06 04:20:38,378] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:38,380] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:38,383] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:38,383] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88 | |
[2023-07-06 04:20:38,383] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,383] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR weight [NNModuleVariable()] | |
[2023-07-06 04:20:38,383] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['self'].input_layernorm.weight (16,) [<DimDynamic.STATIC: 2>] [None] | |
[2023-07-06 04:20:38,383] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:38,384] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:38,384] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable()] | |
[2023-07-06 04:20:38,385] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST input_dtype [GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:38,385] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16)] | |
[2023-07-06 04:20:38,386] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:38,386] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object forward at 0x7fb8cddd7ec0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 83> | |
[2023-07-06 04:20:38,386] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:38,386] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:38,386] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:38,386] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:286 | |
[2023-07-06 04:20:38,386] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,386] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR self_attn [NNModuleVariable()] | |
[2023-07-06 04:20:38,387] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:287 | |
[2023-07-06 04:20:38,387] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [NNModuleVariable()] | |
[2023-07-06 04:20:38,387] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:288 | |
[2023-07-06 04:20:38,387] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST attention_mask [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:38,387] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:289 | |
[2023-07-06 04:20:38,387] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST position_ids [NNModuleVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:38,387] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:290 | |
[2023-07-06 04:20:38,387] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST past_key_value [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:38,387] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:291 | |
[2023-07-06 04:20:38,387] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST output_attentions [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:38,387] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:292 | |
[2023-07-06 04:20:38,387] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ConstantVariable(bool)] | |
[2023-07-06 04:20:38,387] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:286 | |
[2023-07-06 04:20:38,387] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('hidden_states', 'attention_mask', 'position_ids', 'past_key_value', 'output_attentions', 'valid_past_index') [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ConstantVariable(bool), TensorVariable()] | |
[2023-07-06 04:20:38,387] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 6 [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ConstantVariable(bool), TensorVariable(), ConstantVariable(tuple)] | |
[2023-07-06 04:20:38,387] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:38,389] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:38,389] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:38,389] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:38,389] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:38,389] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:38,389] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:38,389] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,389] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:38,390] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function LlamaAttention.forward at 0x7fb8d0c46680>, NNModuleVariable())] | |
[2023-07-06 04:20:38,390] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:38,390] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,390] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:38,390] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,390] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,390] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:38,390] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,391] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,391] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:38,391] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,391] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:38,391] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:38,391] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,391] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:38,391] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:38,391] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:38,391] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,391] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:38,391] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:38,391] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:38,392] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,392] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:38,392] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:38,392] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:38,392] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,392] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:38,392] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:38,392] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:38,392] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:38,392] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:38,392] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:38,392] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function LlamaAttention.forward at 0x7fb8d0c46680>, NNModuleVariable())] | |
[2023-07-06 04:20:38,392] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function LlamaAttention.forward at 0x7fb8d0c46680>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:38,392] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function LlamaAttention.forward at 0x7fb8d0c46680>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:38,392] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function LlamaAttention.forward at 0x7fb8d0c46680>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:38,392] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function LlamaAttention.forward at 0x7fb8d0c46680>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:38,393] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:38,393] torch._dynamo.symbolic_convert: [DEBUG] FAILED INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:38,393] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:38,393] torch._dynamo.symbolic_convert: [DEBUG] break_graph_if_unsupported triggered compile | |
Traceback (most recent call last): | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2187, in inline_call_ | |
sub_locals, closure_cells = func.bind_args(parent, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 187, in bind_args | |
[ | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 188, in <listcomp> | |
wrap(val=arg, source=source) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 62, in wrap_bound_arg | |
return VariableBuilder(tx, source=source, **options)(val) | |
TypeError: VariableBuilder.__init__() got an unexpected keyword argument 'guards' | |
During handling of the above exception, another exception occurred: | |
Traceback (most recent call last): | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 390, in wrapper | |
return inner_fn(self, inst) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 1152, in CALL_FUNCTION_KW | |
self.call_function(fn, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 559, in call_function | |
self.push(fn.call_function(self, args, kwargs)) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/nn_module.py", line 333, in call_function | |
return tx.inline_user_function_return( | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 595, in inline_user_function_return | |
result = InliningInstructionTranslator.inline_call(self, fn, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2132, in inline_call | |
return cls.inline_call_(parent, func, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2229, in inline_call_ | |
tracer.run() | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 708, in run | |
and self.step() | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 668, in step | |
getattr(self, inst.opname)(inst) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 390, in wrapper | |
return inner_fn(self, inst) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 1140, in CALL_FUNCTION_EX | |
self.call_function(fn, argsvars.items, kwargsvars.items) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 559, in call_function | |
self.push(fn.call_function(self, args, kwargs)) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 333, in call_function | |
return super().call_function(tx, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 287, in call_function | |
return super().call_function(tx, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 120, in call_function | |
return tx.inline_user_function_return( | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 595, in inline_user_function_return | |
result = InliningInstructionTranslator.inline_call(self, fn, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2132, in inline_call | |
return cls.inline_call_(parent, func, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2190, in inline_call_ | |
raise ArgsMismatchError( | |
torch._dynamo.exc.ArgsMismatchError: VariableBuilder.__init__() got an unexpected keyword argument 'guards'. | |
func = 'forward' /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:190, args = [<class 'trfs_fast.llama.LlamaAttention'>], kwargs = OrderedDict([('hidden_states', TensorVariable()), ('attention_mask', TensorVariable()), ('position_ids', TensorVariable()), ('past_key_value', TensorVariable()), ('output_attentions', ConstantVariable(bool)), ('valid_past_index', TensorVariable())]) | |
[2023-07-06 04:20:38,393] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:38,393] torch._dynamo.output_graph: [DEBUG] COMPILING GRAPH due to GraphCompileReason(reason="VariableBuilder.__init__() got an unexpected keyword argument 'guards'.\n func = 'forward' /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:190, args = [<class 'trfs_fast.llama.LlamaAttention'>], kwargs = OrderedDict([('hidden_states', TensorVariable()), ('attention_mask', TensorVariable()), ('position_ids', TensorVariable()), ('past_key_value', TensorVariable()), ('output_attentions', ConstantVariable(bool)), ('valid_past_index', TensorVariable())])", user_stack=[<FrameSummary file /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py, line 286 in forward>, <FrameSummary file /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py, line 1514 in _call_impl>], graph_break=True) | |
[2023-07-06 04:20:38,394] torch._dynamo.output_graph: [DEBUG] REMOVE UNUSED GRAPHARG L['attention_mask'] | |
[2023-07-06 04:20:38,395] torch._dynamo.output_graph: [DEBUG] REMOVE UNUSED GRAPHARG L['position_ids'] | |
[2023-07-06 04:20:38,395] torch._dynamo.output_graph: [DEBUG] REMOVE UNUSED GRAPHARG L['past_key_value'] | |
[2023-07-06 04:20:38,395] torch._dynamo.output_graph: [DEBUG] REMOVE UNUSED GRAPHARG L['valid_past_index'] | |
[2023-07-06 04:20:38,395] torch._dynamo.output_graph.__graph_code: [DEBUG] TRACED GRAPH | |
===== __compiled_fn_1 ===== | |
<eval_with_key>.0 class GraphModule(torch.nn.Module): | |
def forward(self, L_hidden_states_ : torch.Tensor): | |
l_hidden_states_ = L_hidden_states_ | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85, code: variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) | |
to = l_hidden_states_.to(torch.float32) | |
pow_1 = to.pow(2); to = None | |
mean = pow_1.mean(-1, keepdim = True); pow_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86, code: hidden_states = hidden_states * torch.rsqrt(variance + self.variance_epsilon) | |
add = mean + 1e-06; mean = None | |
rsqrt = torch.rsqrt(add); add = None | |
mul = l_hidden_states_ * rsqrt; l_hidden_states_ = rsqrt = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88, code: return (self.weight * hidden_states).to(input_dtype) | |
l__self___input_layernorm_weight = self.L__self___input_layernorm_weight | |
mul_1 = l__self___input_layernorm_weight * mul; l__self___input_layernorm_weight = mul = None | |
to_1 = mul_1.to(torch.float16); mul_1 = None | |
return (to_1,) | |
[2023-07-06 04:20:38,396] torch._dynamo.output_graph.__graph: [DEBUG] TRACED GRAPH | |
__compiled_fn_1 <eval_with_key>.0 opcode name target args kwargs | |
------------- -------------------------------- -------------------------------------------------------- --------------------------------------- ----------------- | |
placeholder l_hidden_states_ L_hidden_states_ () {} | |
call_method to to (l_hidden_states_, torch.float32) {} | |
call_method pow_1 pow (to, 2) {} | |
call_method mean mean (pow_1, -1) {'keepdim': True} | |
call_function add <built-in function add> (mean, 1e-06) {} | |
call_function rsqrt <built-in method rsqrt of type object at 0x7fb940469580> (add,) {} | |
call_function mul <built-in function mul> (l_hidden_states_, rsqrt) {} | |
get_attr l__self___input_layernorm_weight L__self___input_layernorm_weight () {} | |
call_function mul_1 <built-in function mul> (l__self___input_layernorm_weight, mul) {} | |
call_method to_1 to (mul_1, torch.float16) {} | |
output output output ((to_1,),) {} | |
[2023-07-06 04:20:38,401] torch._dynamo.output_graph: [INFO] Step 2: calling compiler function inductor | |
Input hash: e4edb9de | |
Cache preallocation: True | |
Warming up (2 runs)... | |
[2023-07-06 04:20:38,710] torch._functorch.aot_autograd.__aot_graphs: [INFO] TRACED GRAPH | |
===== Forward graph 0 ===== | |
<eval_with_key>.4 class <lambda>(torch.nn.Module): | |
def forward(self, arg0_1: f16[16], arg1_1: f16[1, 1000, 16]): | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85, code: variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) | |
convert_element_type: f32[1, 1000, 16] = torch.ops.prims.convert_element_type.default(arg1_1, torch.float32) | |
pow_1: f32[1, 1000, 16] = torch.ops.aten.pow.Tensor_Scalar(convert_element_type, 2); convert_element_type = None | |
mean: f32[1, 1000, 1] = torch.ops.aten.mean.dim(pow_1, [-1], True); pow_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86, code: hidden_states = hidden_states * torch.rsqrt(variance + self.variance_epsilon) | |
add: f32[1, 1000, 1] = torch.ops.aten.add.Tensor(mean, 1e-06); mean = None | |
rsqrt: f32[1, 1000, 1] = torch.ops.aten.rsqrt.default(add); add = None | |
mul: f32[1, 1000, 16] = torch.ops.aten.mul.Tensor(arg1_1, rsqrt); arg1_1 = rsqrt = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88, code: return (self.weight * hidden_states).to(input_dtype) | |
mul_1: f32[1, 1000, 16] = torch.ops.aten.mul.Tensor(arg0_1, mul); arg0_1 = mul = None | |
convert_element_type_1: f16[1, 1000, 16] = torch.ops.prims.convert_element_type.default(mul_1, torch.float16); mul_1 = None | |
return (convert_element_type_1,) | |
[2023-07-06 04:20:40,125] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 0 | |
[2023-07-06 04:20:40,183] torch._inductor.graph: [DEBUG] Force channels last inputs for 0 conv for the current graph with id 0 | |
[2023-07-06 04:20:41,671] torch._inductor.scheduler: [DEBUG] remove_buffer('buf0') | |
[2023-07-06 04:20:42,300] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 4 configs | |
[2023-07-06 04:20:42,300] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 1, num_warps: 2, num_stages: 1 | |
[2023-07-06 04:20:42,300] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 8, num_warps: 2, num_stages: 1 | |
[2023-07-06 04:20:42,300] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 32, num_warps: 4, num_stages: 1 | |
[2023-07-06 04:20:42,300] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 128, num_warps: 8, num_stages: 1 | |
[2023-07-06 04:20:43,598] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 4 configs | |
[2023-07-06 04:20:43,598] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 1, num_warps: 2, num_stages: 1 | |
[2023-07-06 04:20:43,598] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 8, num_warps: 2, num_stages: 1 | |
[2023-07-06 04:20:43,598] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 32, num_warps: 4, num_stages: 1 | |
[2023-07-06 04:20:43,598] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 128, num_warps: 8, num_stages: 1 | |
[2023-07-06 04:20:43,720] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_ubuntu/tn/ctnjyv37bz6fvtmwelga5764xq272722mvbuxkb4qqgd3syt7hwl.py | |
[2023-07-06 04:20:43,725] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 0 | |
[2023-07-06 04:20:43,727] torch._dynamo.output_graph: [INFO] Step 2: done compiler function inductor | |
[2023-07-06 04:20:43,737] torch._dynamo.convert_frame.__guards: [DEBUG] GUARDS: | |
___check_obj_id(L['self'], 140431662763072) | |
hasattr(L['position_ids'], '_dynamo_dynamic_indices') == False | |
___check_type_id(L['hidden_states'], 77765680) | |
hasattr(L['hidden_states'], '_dynamo_dynamic_indices') == False | |
hasattr(L['attention_mask'], '_dynamo_dynamic_indices') == False | |
hasattr(L['past_key_value'], '_dynamo_dynamic_indices') == False | |
hasattr(L['valid_past_index'], '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['output_attentions'], 7698112) | |
not ___is_grad_enabled() | |
not ___are_deterministic_algorithms_enabled() | |
utils_device.CURRENT_DEVICE == None | |
str(G['torch'].float32) == 'torch.float32' | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks.keys()) == set() | |
[2023-07-06 04:20:43,737] torch._dynamo.eval_frame: [DEBUG] skipping _fn /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py | |
[2023-07-06 04:20:43,738] torch._dynamo.eval_frame: [DEBUG] skipping nothing /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py | |
[2023-07-06 04:20:43,844] torch._inductor.cudagraph_trees: [DEBUG] Running warmup of function 0 | |
[2023-07-06 04:20:44,100] torch._inductor.triton_heuristics: [DEBUG] Benchmark all input configs get: | |
[2023-07-06 04:20:44,100] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 1, num_warps: 2, num_stages: 1: 0.005120, nreg 16, nspill 0, #shared-mem 256 | |
[2023-07-06 04:20:44,100] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 8, num_warps: 2, num_stages: 1: 0.005120, nreg 16, nspill 0, #shared-mem 256 | |
[2023-07-06 04:20:44,100] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 32, num_warps: 4, num_stages: 1: 0.005120, nreg 24, nspill 0, #shared-mem 512 | |
[2023-07-06 04:20:44,100] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 128, num_warps: 8, num_stages: 1: 0.005120, nreg 26, nspill 0, #shared-mem 1024 | |
[2023-07-06 04:20:44,100] torch._inductor.triton_heuristics: [DEBUG] Save heuristic tuning result to /tmp/torchinductor_ubuntu/ms/cmsiodjdcnbq2hee6trx6wljayuvfvkl3gqgzjhqci2lua23fz6n.best_config | |
[2023-07-06 04:20:44,100] torch._dynamo.eval_frame: [DEBUG] skipping __exit__ /home/ubuntu/miniconda3/lib/python3.10/contextlib.py | |
[2023-07-06 04:20:44,100] torch._dynamo.eval_frame: [DEBUG] skipping __exit__ /home/ubuntu/miniconda3/lib/python3.10/contextlib.py | |
[2023-07-06 04:20:44,101] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo start tracing forward /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:190 | |
[2023-07-06 04:20:44,102] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['hidden_states'] (1, 1000, 16) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None] | |
[2023-07-06 04:20:44,103] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['attention_mask'] (1, 1003) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None] | |
[2023-07-06 04:20:44,104] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['position_ids'] (1, 1000) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None] | |
[2023-07-06 04:20:44,105] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['past_key_value'] (2, 1, 4, 1003, 4) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None, None, None] | |
[2023-07-06 04:20:44,105] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['valid_past_index'] () [] [] | |
[2023-07-06 04:20:44,106] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:199 | |
[2023-07-06 04:20:44,106] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST output_attentions [] | |
[2023-07-06 04:20:44,106] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST True [ConstantVariable(bool)] | |
[2023-07-06 04:20:44,106] torch._dynamo.symbolic_convert: [DEBUG] TRACE IS_OP 0 [ConstantVariable(bool), ConstantVariable(bool)] | |
[2023-07-06 04:20:44,106] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(bool)] | |
[2023-07-06 04:20:44,106] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:202 | |
[2023-07-06 04:20:44,106] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:44,106] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR size [TensorVariable()] | |
[2023-07-06 04:20:44,107] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [GetAttrVariable(TensorVariable(), size)] | |
[2023-07-06 04:20:44,107] torch._dynamo.symbolic_convert: [DEBUG] TRACE UNPACK_SEQUENCE 3 [SizeVariable()] | |
[2023-07-06 04:20:44,107] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST bsz [ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:44,107] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST q_len [ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:44,107] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST _ [ConstantVariable(int)] | |
[2023-07-06 04:20:44,107] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:205 | |
[2023-07-06 04:20:44,107] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:44,107] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR qkv_proj [NNModuleVariable()] | |
[2023-07-06 04:20:44,108] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [NNModuleVariable()] | |
[2023-07-06 04:20:44,108] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,111] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR view [TensorVariable()] | |
[2023-07-06 04:20:44,111] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST bsz [GetAttrVariable(TensorVariable(), view)] | |
[2023-07-06 04:20:44,111] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST q_len [GetAttrVariable(TensorVariable(), view), ConstantVariable(int)] | |
[2023-07-06 04:20:44,111] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 3 [GetAttrVariable(TensorVariable(), view), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:44,111] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [GetAttrVariable(TensorVariable(), view), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:44,111] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR num_heads [GetAttrVariable(TensorVariable(), view), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), NNModuleVariable()] | |
[2023-07-06 04:20:44,112] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [GetAttrVariable(TensorVariable(), view), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:44,112] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR head_dim [GetAttrVariable(TensorVariable(), view), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), NNModuleVariable()] | |
[2023-07-06 04:20:44,112] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 5 [GetAttrVariable(TensorVariable(), view), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:44,113] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR permute [TensorVariable()] | |
[2023-07-06 04:20:44,113] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [GetAttrVariable(TensorVariable(), permute)] | |
[2023-07-06 04:20:44,113] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [GetAttrVariable(TensorVariable(), permute), ConstantVariable(int)] | |
[2023-07-06 04:20:44,113] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 3 [GetAttrVariable(TensorVariable(), permute), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:44,113] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [GetAttrVariable(TensorVariable(), permute), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:44,113] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 4 [GetAttrVariable(TensorVariable(), permute), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:44,113] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 5 [GetAttrVariable(TensorVariable(), permute), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:44,114] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST query_key_value_states [TensorVariable()] | |
[2023-07-06 04:20:44,114] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:207 | |
[2023-07-06 04:20:44,114] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST query_key_value_states [] | |
[2023-07-06 04:20:44,114] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [TensorVariable()] | |
[2023-07-06 04:20:44,114] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:44,115] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST query_states [TensorVariable()] | |
[2023-07-06 04:20:44,115] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:208 | |
[2023-07-06 04:20:44,115] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST query_key_value_states [] | |
[2023-07-06 04:20:44,115] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [TensorVariable()] | |
[2023-07-06 04:20:44,115] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:44,115] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_SLICE 2 [TensorVariable(), ConstantVariable(int), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:44,116] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), SliceVariable()] | |
[2023-07-06 04:20:44,117] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST key_value_states [TensorVariable()] | |
[2023-07-06 04:20:44,117] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:211 | |
[2023-07-06 04:20:44,117] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:44,117] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR rotary_emb [NNModuleVariable()] | |
[2023-07-06 04:20:44,117] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_value_states [NNModuleVariable()] | |
[2023-07-06 04:20:44,117] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,117] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST q_len [NNModuleVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,117] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [NNModuleVariable(), TensorVariable(), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:44,118] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('seq_len',) [NNModuleVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,118] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 2 [NNModuleVariable(), TensorVariable(), TensorVariable(), ConstantVariable(tuple)] | |
[2023-07-06 04:20:44,119] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:44,120] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:44,120] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:44,120] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:44,120] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:44,121] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:44,121] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:44,121] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:44,121] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:44,121] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function LlamaRotaryEmbedding.forward at 0x7fb8d0c46290>, NNModuleVariable())] | |
[2023-07-06 04:20:44,121] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:44,121] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:44,121] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:44,122] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:44,122] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:44,122] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:44,122] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:44,122] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:44,122] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:44,122] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:44,122] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:44,123] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:44,123] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:44,123] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:44,123] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:44,123] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:44,123] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:44,123] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:44,123] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:44,123] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:44,123] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:44,123] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:44,123] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:44,123] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:44,124] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:44,124] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:44,124] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:44,124] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:44,124] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:44,124] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:44,124] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:44,124] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function LlamaRotaryEmbedding.forward at 0x7fb8d0c46290>, NNModuleVariable())] | |
[2023-07-06 04:20:44,124] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function LlamaRotaryEmbedding.forward at 0x7fb8d0c46290>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:44,124] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function LlamaRotaryEmbedding.forward at 0x7fb8d0c46290>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:44,124] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function LlamaRotaryEmbedding.forward at 0x7fb8d0c46290>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:44,124] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function LlamaRotaryEmbedding.forward at 0x7fb8d0c46290>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:44,124] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object forward at 0x7fb8cddf80e0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 107> | |
[2023-07-06 04:20:44,125] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:122 | |
[2023-07-06 04:20:44,125] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:44,125] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR cos_cached [NNModuleVariable()] | |
[2023-07-06 04:20:44,125] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['self'].rotary_emb.cos_cached (1, 1, 2048, 4) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None, None] | |
[2023-07-06 04:20:44,125] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable()] | |
[2023-07-06 04:20:44,126] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:44,126] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR dtype [GetAttrVariable(TensorVariable(), to), TensorVariable()] | |
[2023-07-06 04:20:44,126] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('dtype',) [GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16)] | |
[2023-07-06 04:20:44,126] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 1 [GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16), ConstantVariable(tuple)] | |
[2023-07-06 04:20:44,127] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:123 | |
[2023-07-06 04:20:44,127] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [TensorVariable()] | |
[2023-07-06 04:20:44,127] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR sin_cached [TensorVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:44,128] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['self'].rotary_emb.sin_cached (1, 1, 2048, 4) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None, None] | |
[2023-07-06 04:20:44,128] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,129] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TensorVariable(), GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:44,129] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR dtype [TensorVariable(), GetAttrVariable(TensorVariable(), to), TensorVariable()] | |
[2023-07-06 04:20:44,129] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('dtype',) [TensorVariable(), GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16)] | |
[2023-07-06 04:20:44,129] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 1 [TensorVariable(), GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16), ConstantVariable(tuple)] | |
[2023-07-06 04:20:44,130] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:121 | |
[2023-07-06 04:20:44,130] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_TUPLE 2 [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,130] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TupleVariable()] | |
[2023-07-06 04:20:44,130] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object forward at 0x7fb8cddf80e0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 107> | |
[2023-07-06 04:20:44,130] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TupleVariable()] | |
[2023-07-06 04:20:44,131] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:44,131] torch._dynamo.symbolic_convert: [DEBUG] TRACE UNPACK_SEQUENCE 2 [TupleVariable()] | |
[2023-07-06 04:20:44,131] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST cos [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,131] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST sin [TensorVariable()] | |
[2023-07-06 04:20:44,131] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:212 | |
[2023-07-06 04:20:44,131] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL apply_rotary_pos_emb_opt [] | |
[2023-07-06 04:20:44,131] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST query_states [UserFunctionVariable()] | |
[2023-07-06 04:20:44,131] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_value_states [UserFunctionVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,131] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [UserFunctionVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,131] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [UserFunctionVariable(), TensorVariable(), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:44,132] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST cos [UserFunctionVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,132] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST sin [UserFunctionVariable(), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,132] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST position_ids [UserFunctionVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,132] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 5 [UserFunctionVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,132] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object apply_rotary_pos_emb_opt at 0x7fb8cddf82f0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 135> | |
[2023-07-06 04:20:44,132] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:139 | |
[2023-07-06 04:20:44,132] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST cos [] | |
[2023-07-06 04:20:44,132] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR squeeze [TensorVariable()] | |
[2023-07-06 04:20:44,133] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [GetAttrVariable(TensorVariable(), squeeze)] | |
[2023-07-06 04:20:44,133] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), squeeze), ConstantVariable(int)] | |
[2023-07-06 04:20:44,134] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR squeeze [TensorVariable()] | |
[2023-07-06 04:20:44,134] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [GetAttrVariable(TensorVariable(), squeeze)] | |
[2023-07-06 04:20:44,134] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), squeeze), ConstantVariable(int)] | |
[2023-07-06 04:20:44,135] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST cos [TensorVariable()] | |
[2023-07-06 04:20:44,135] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:140 | |
[2023-07-06 04:20:44,135] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST sin [] | |
[2023-07-06 04:20:44,135] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR squeeze [TensorVariable()] | |
[2023-07-06 04:20:44,135] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [GetAttrVariable(TensorVariable(), squeeze)] | |
[2023-07-06 04:20:44,135] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), squeeze), ConstantVariable(int)] | |
[2023-07-06 04:20:44,136] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR squeeze [TensorVariable()] | |
[2023-07-06 04:20:44,136] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [GetAttrVariable(TensorVariable(), squeeze)] | |
[2023-07-06 04:20:44,136] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), squeeze), ConstantVariable(int)] | |
[2023-07-06 04:20:44,137] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST sin [TensorVariable()] | |
[2023-07-06 04:20:44,137] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:141 | |
[2023-07-06 04:20:44,137] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST cos [] | |
[2023-07-06 04:20:44,137] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST position_ids [TensorVariable()] | |
[2023-07-06 04:20:44,137] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,138] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR unsqueeze [TensorVariable()] | |
[2023-07-06 04:20:44,139] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [GetAttrVariable(TensorVariable(), unsqueeze)] | |
[2023-07-06 04:20:44,139] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), unsqueeze), ConstantVariable(int)] | |
[2023-07-06 04:20:44,139] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST cos [TensorVariable()] | |
[2023-07-06 04:20:44,140] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:142 | |
[2023-07-06 04:20:44,140] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST sin [] | |
[2023-07-06 04:20:44,140] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST position_ids [TensorVariable()] | |
[2023-07-06 04:20:44,140] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,141] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR unsqueeze [TensorVariable()] | |
[2023-07-06 04:20:44,141] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [GetAttrVariable(TensorVariable(), unsqueeze)] | |
[2023-07-06 04:20:44,141] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), unsqueeze), ConstantVariable(int)] | |
[2023-07-06 04:20:44,142] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST sin [TensorVariable()] | |
[2023-07-06 04:20:44,142] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:143 | |
[2023-07-06 04:20:44,142] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST q [] | |
[2023-07-06 04:20:44,142] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST cos [TensorVariable()] | |
[2023-07-06 04:20:44,142] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,143] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL rotate_half [TensorVariable()] | |
[2023-07-06 04:20:44,143] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST q [TensorVariable(), UserFunctionVariable()] | |
[2023-07-06 04:20:44,143] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [TensorVariable(), UserFunctionVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,143] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object rotate_half at 0x7fb8cddf8240, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 127> | |
[2023-07-06 04:20:44,144] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:129 | |
[2023-07-06 04:20:44,144] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:44,144] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR roll [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:44,144] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TorchVariable(<built-in method roll of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:44,144] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:44,144] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR shape [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,144] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ShapeVariable()] | |
[2023-07-06 04:20:44,145] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ShapeVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:44,145] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:44,145] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_FLOOR_DIVIDE None [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:44,145] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:44,145] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('shifts', 'dims') [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:44,145] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 3 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int), ConstantVariable(int), ConstantVariable(tuple)] | |
[2023-07-06 04:20:44,146] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST x [TensorVariable()] | |
[2023-07-06 04:20:44,146] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:130 | |
[2023-07-06 04:20:44,147] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [] | |
[2023-07-06 04:20:44,147] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST Ellipsis [TensorVariable()] | |
[2023-07-06 04:20:44,147] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [TensorVariable(), ConstantVariable(ellipsis)] | |
[2023-07-06 04:20:44,147] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:44,147] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR shape [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), TensorVariable()] | |
[2023-07-06 04:20:44,147] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ShapeVariable()] | |
[2023-07-06 04:20:44,147] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ShapeVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:44,147] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ConstantVariable(int)] | |
[2023-07-06 04:20:44,147] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_FLOOR_DIVIDE None [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:44,147] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_SLICE 2 [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ConstantVariable(int)] | |
[2023-07-06 04:20:44,148] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_TUPLE 2 [TensorVariable(), ConstantVariable(ellipsis), SliceVariable()] | |
[2023-07-06 04:20:44,148] torch._dynamo.symbolic_convert: [DEBUG] TRACE DUP_TOP_TWO None [TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:44,148] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), TupleVariable(), TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:44,149] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TensorVariable(), TupleVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,149] torch._dynamo.symbolic_convert: [DEBUG] TRACE INPLACE_MULTIPLY None [TensorVariable(), TupleVariable(), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:44,149] torch._dynamo.symbolic_convert: [DEBUG] TRACE ROT_THREE None [TensorVariable(), TupleVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,150] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_SUBSCR None [TensorVariable(), TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:44,150] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:131 | |
[2023-07-06 04:20:44,150] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [] | |
[2023-07-06 04:20:44,150] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:44,150] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object rotate_half at 0x7fb8cddf8240, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 127> | |
[2023-07-06 04:20:44,150] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST sin [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,150] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,151] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,152] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST q_embed [TensorVariable()] | |
[2023-07-06 04:20:44,152] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:144 | |
[2023-07-06 04:20:44,152] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_states [] | |
[2023-07-06 04:20:44,152] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR copy_ [TensorVariable()] | |
[2023-07-06 04:20:44,152] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_states [GetAttrVariable(TensorVariable(), copy_)] | |
[2023-07-06 04:20:44,152] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST cos [GetAttrVariable(TensorVariable(), copy_), TensorVariable()] | |
[2023-07-06 04:20:44,152] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [GetAttrVariable(TensorVariable(), copy_), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,153] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL rotate_half [GetAttrVariable(TensorVariable(), copy_), TensorVariable()] | |
[2023-07-06 04:20:44,153] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_states [GetAttrVariable(TensorVariable(), copy_), TensorVariable(), UserFunctionVariable()] | |
[2023-07-06 04:20:44,153] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), copy_), TensorVariable(), UserFunctionVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,153] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object rotate_half at 0x7fb8cddf8240, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 127> | |
[2023-07-06 04:20:44,154] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:129 | |
[2023-07-06 04:20:44,154] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:44,154] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR roll [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:44,154] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TorchVariable(<built-in method roll of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:44,154] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:44,154] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR shape [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,155] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ShapeVariable()] | |
[2023-07-06 04:20:44,155] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ShapeVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:44,155] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:44,155] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_FLOOR_DIVIDE None [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:44,155] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:44,155] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('shifts', 'dims') [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:44,155] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 3 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int), ConstantVariable(int), ConstantVariable(tuple)] | |
[2023-07-06 04:20:44,156] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST x [TensorVariable()] | |
[2023-07-06 04:20:44,156] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:130 | |
[2023-07-06 04:20:44,156] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [] | |
[2023-07-06 04:20:44,156] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST Ellipsis [TensorVariable()] | |
[2023-07-06 04:20:44,156] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [TensorVariable(), ConstantVariable(ellipsis)] | |
[2023-07-06 04:20:44,156] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:44,156] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR shape [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), TensorVariable()] | |
[2023-07-06 04:20:44,157] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ShapeVariable()] | |
[2023-07-06 04:20:44,157] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ShapeVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:44,157] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ConstantVariable(int)] | |
[2023-07-06 04:20:44,157] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_FLOOR_DIVIDE None [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:44,157] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_SLICE 2 [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ConstantVariable(int)] | |
[2023-07-06 04:20:44,157] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_TUPLE 2 [TensorVariable(), ConstantVariable(ellipsis), SliceVariable()] | |
[2023-07-06 04:20:44,157] torch._dynamo.symbolic_convert: [DEBUG] TRACE DUP_TOP_TWO None [TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:44,157] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), TupleVariable(), TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:44,158] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TensorVariable(), TupleVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,158] torch._dynamo.symbolic_convert: [DEBUG] TRACE INPLACE_MULTIPLY None [TensorVariable(), TupleVariable(), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:44,159] torch._dynamo.symbolic_convert: [DEBUG] TRACE ROT_THREE None [TensorVariable(), TupleVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,159] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_SUBSCR None [TensorVariable(), TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:44,159] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:131 | |
[2023-07-06 04:20:44,159] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [] | |
[2023-07-06 04:20:44,159] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:44,159] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object rotate_half at 0x7fb8cddf8240, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 127> | |
[2023-07-06 04:20:44,159] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST sin [GetAttrVariable(TensorVariable(), copy_), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,159] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [GetAttrVariable(TensorVariable(), copy_), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,160] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [GetAttrVariable(TensorVariable(), copy_), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,161] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), copy_), TensorVariable()] | |
[2023-07-06 04:20:44,162] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_TOP None [TensorVariable()] | |
[2023-07-06 04:20:44,162] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:145 | |
[2023-07-06 04:20:44,162] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST q_embed [] | |
[2023-07-06 04:20:44,162] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:44,162] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object apply_rotary_pos_emb_opt at 0x7fb8cddf82f0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 135> | |
[2023-07-06 04:20:44,162] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST query_states [TensorVariable()] | |
[2023-07-06 04:20:44,162] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:215 | |
[2023-07-06 04:20:44,162] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [] | |
[2023-07-06 04:20:44,162] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST past_kv_slice_start [TensorVariable()] | |
[2023-07-06 04:20:44,162] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:216 | |
[2023-07-06 04:20:44,162] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:44,162] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR eq [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:44,163] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [TorchVariable(<built-in method eq of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:44,163] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [TorchVariable(<built-in method eq of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:44,163] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 2 [TorchVariable(<built-in method eq of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:44,164] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR int [TensorVariable()] | |
[2023-07-06 04:20:44,164] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [GetAttrVariable(TensorVariable(), int)] | |
[2023-07-06 04:20:44,165] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST q_len [TensorVariable()] | |
[2023-07-06 04:20:44,165] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:44,166] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [TensorVariable()] | |
[2023-07-06 04:20:44,166] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR ne [TensorVariable(), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:44,166] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [TensorVariable(), TorchVariable(<built-in method ne of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:44,166] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [TensorVariable(), TorchVariable(<built-in method ne of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:44,166] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 2 [TensorVariable(), TorchVariable(<built-in method ne of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:44,167] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR int [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,168] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TensorVariable(), GetAttrVariable(TensorVariable(), int)] | |
[2023-07-06 04:20:44,169] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,169] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,169] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TensorVariable(), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:44,169] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,170] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,171] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST past_kv_slice_end [TensorVariable()] | |
[2023-07-06 04:20:44,171] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:217 | |
[2023-07-06 04:20:44,171] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:44,171] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR eq [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:44,172] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [TorchVariable(<built-in method eq of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:44,172] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [TorchVariable(<built-in method eq of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:44,172] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 2 [TorchVariable(<built-in method eq of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:44,172] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR int [TensorVariable()] | |
[2023-07-06 04:20:44,173] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [GetAttrVariable(TensorVariable(), int)] | |
[2023-07-06 04:20:44,174] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_value_states [TensorVariable()] | |
[2023-07-06 04:20:44,174] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR shape [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,174] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -2 [TensorVariable(), ShapeVariable()] | |
[2023-07-06 04:20:44,174] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), ShapeVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:44,174] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:44,175] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [TensorVariable()] | |
[2023-07-06 04:20:44,175] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR ne [TensorVariable(), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:44,175] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [TensorVariable(), TorchVariable(<built-in method ne of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:44,176] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [TensorVariable(), TorchVariable(<built-in method ne of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:44,176] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 2 [TensorVariable(), TorchVariable(<built-in method ne of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:44,176] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR int [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,177] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TensorVariable(), GetAttrVariable(TensorVariable(), int)] | |
[2023-07-06 04:20:44,178] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST past_kv_slice_end [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,178] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,178] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,179] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST past_state_slice_end [TensorVariable()] | |
[2023-07-06 04:20:44,179] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:218 | |
[2023-07-06 04:20:44,179] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_value_states [] | |
[2023-07-06 04:20:44,179] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST past_key_value [TensorVariable()] | |
[2023-07-06 04:20:44,179] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST Ellipsis [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,179] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST past_kv_slice_start [TensorVariable(), TensorVariable(), ConstantVariable(ellipsis)] | |
[2023-07-06 04:20:44,179] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST past_kv_slice_end [TensorVariable(), TensorVariable(), ConstantVariable(ellipsis), TensorVariable()] | |
[2023-07-06 04:20:44,180] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_SLICE 2 [TensorVariable(), TensorVariable(), ConstantVariable(ellipsis), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:44,180] torch._dynamo.symbolic_convert: [DEBUG] step triggered compile | |
Traceback (most recent call last): | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 668, in step | |
getattr(self, inst.opname)(inst) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 1251, in BUILD_SLICE | |
SliceVariable( | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/base.py", line 27, in __call__ | |
obj = type.__call__(cls, *args, **kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/lists.py", line 616, in __init__ | |
unimplemented("Dynamic slicing on data-dependent value is not supported") | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/exc.py", line 140, in unimplemented | |
raise Unsupported(msg) | |
torch._dynamo.exc.Unsupported: Dynamic slicing on data-dependent value is not supported | |
[2023-07-06 04:20:44,181] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:44,182] torch._dynamo.output_graph: [DEBUG] COMPILING GRAPH due to GraphCompileReason(reason='step_unsupported', user_stack=[<FrameSummary file /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py, line 218 in forward>], graph_break=True) | |
[2023-07-06 04:20:44,183] torch._dynamo.output_graph: [DEBUG] REMOVE UNUSED GRAPHARG L['attention_mask'] | |
[2023-07-06 04:20:44,183] torch._dynamo.output_graph: [DEBUG] REMOVE UNUSED GRAPHARG L['past_key_value'] | |
[2023-07-06 04:20:44,185] torch._dynamo.output_graph.__graph_code: [DEBUG] TRACED GRAPH | |
===== __compiled_fn_3 ===== | |
<eval_with_key>.111 class GraphModule(torch.nn.Module): | |
def forward(self, L_hidden_states_ : torch.Tensor, L_position_ids_ : torch.Tensor, L_valid_past_index_ : torch.Tensor): | |
l_hidden_states_ = L_hidden_states_ | |
l_position_ids_ = L_position_ids_ | |
l_valid_past_index_ = L_valid_past_index_ | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:205, code: query_key_value_states = self.qkv_proj(hidden_states).view(bsz, q_len, 3, self.num_heads, self.head_dim).permute(2, 0, 3, 1, 4) | |
l__self___qkv_proj = self.L__self___qkv_proj(l_hidden_states_); l_hidden_states_ = None | |
view = l__self___qkv_proj.view(1, 1000, 3, 4, 4); l__self___qkv_proj = None | |
permute = view.permute(2, 0, 3, 1, 4); view = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:207, code: query_states = query_key_value_states[0] | |
getitem = permute[0] | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:208, code: key_value_states = query_key_value_states[1:] | |
getitem_1 = permute[slice(1, None, None)]; permute = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:211, code: cos, sin = self.rotary_emb(key_value_states, seq_len=valid_past_index + q_len) | |
add = l_valid_past_index_ + 1000 | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:122, code: self.cos_cached.to(dtype=x.dtype), | |
l__self___rotary_emb_cos_cached = self.L__self___rotary_emb_cos_cached | |
to = l__self___rotary_emb_cos_cached.to(dtype = torch.float16); l__self___rotary_emb_cos_cached = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:123, code: self.sin_cached.to(dtype=x.dtype), | |
l__self___rotary_emb_sin_cached = self.L__self___rotary_emb_sin_cached | |
to_1 = l__self___rotary_emb_sin_cached.to(dtype = torch.float16); l__self___rotary_emb_sin_cached = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:212, code: query_states = apply_rotary_pos_emb_opt(query_states, key_value_states[0], cos, sin, position_ids) | |
getitem_2 = getitem_1[0] | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:139, code: cos = cos.squeeze(1).squeeze(0) # [seq_len, dim] | |
squeeze = to.squeeze(1); to = None | |
squeeze_1 = squeeze.squeeze(0); squeeze = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:140, code: sin = sin.squeeze(1).squeeze(0) # [seq_len, dim] | |
squeeze_2 = to_1.squeeze(1); to_1 = None | |
squeeze_3 = squeeze_2.squeeze(0); squeeze_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:141, code: cos = cos[position_ids].unsqueeze(1) # [bs, 1, seq_len, dim] | |
getitem_3 = squeeze_1[l_position_ids_]; squeeze_1 = None | |
unsqueeze = getitem_3.unsqueeze(1); getitem_3 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:142, code: sin = sin[position_ids].unsqueeze(1) # [bs, 1, seq_len, dim] | |
getitem_4 = squeeze_3[l_position_ids_]; squeeze_3 = l_position_ids_ = None | |
unsqueeze_1 = getitem_4.unsqueeze(1); getitem_4 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:143, code: q_embed = (q * cos) + (rotate_half(q) * sin) | |
mul = getitem * unsqueeze | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:129, code: x = torch.roll(x, shifts=x.shape[-1] // 2, dims=-1) | |
roll = torch.roll(getitem, shifts = 2, dims = -1); getitem = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:130, code: x[..., : x.shape[-1] // 2] *= -1 | |
getitem_5 = roll[(Ellipsis, slice(None, 2, None))] | |
getitem_5 *= -1; imul = getitem_5; getitem_5 = None | |
roll[(Ellipsis, slice(None, 2, None))] = imul; setitem = roll; imul = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:143, code: q_embed = (q * cos) + (rotate_half(q) * sin) | |
mul_1 = roll * unsqueeze_1; roll = None | |
add_1 = mul + mul_1; mul = mul_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:144, code: key_states.copy_((key_states * cos) + (rotate_half(key_states) * sin)) | |
mul_2 = getitem_2 * unsqueeze; unsqueeze = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:129, code: x = torch.roll(x, shifts=x.shape[-1] // 2, dims=-1) | |
roll_1 = torch.roll(getitem_2, shifts = 2, dims = -1) | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:130, code: x[..., : x.shape[-1] // 2] *= -1 | |
getitem_6 = roll_1[(Ellipsis, slice(None, 2, None))] | |
getitem_6 *= -1; imul_1 = getitem_6; getitem_6 = None | |
roll_1[(Ellipsis, slice(None, 2, None))] = imul_1; setitem_1 = roll_1; imul_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:144, code: key_states.copy_((key_states * cos) + (rotate_half(key_states) * sin)) | |
mul_3 = roll_1 * unsqueeze_1; roll_1 = unsqueeze_1 = None | |
add_2 = mul_2 + mul_3; mul_2 = mul_3 = None | |
copy_ = getitem_2.copy_(add_2); getitem_2 = add_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:216, code: past_kv_slice_end = torch.eq(valid_past_index, 0).int() * q_len + torch.ne(valid_past_index, 0).int() * (valid_past_index + 1) | |
eq = torch.eq(l_valid_past_index_, 0) | |
int_1 = eq.int(); eq = None | |
mul_4 = int_1 * 1000; int_1 = None | |
ne = torch.ne(l_valid_past_index_, 0) | |
int_2 = ne.int(); ne = None | |
add_3 = l_valid_past_index_ + 1 | |
mul_5 = int_2 * add_3; int_2 = add_3 = None | |
add_4 = mul_4 + mul_5; mul_4 = mul_5 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:217, code: past_state_slice_end = torch.eq(valid_past_index, 0).int() * key_value_states.shape[-2] + torch.ne(valid_past_index, 0).int() * (past_kv_slice_end) | |
eq_1 = torch.eq(l_valid_past_index_, 0) | |
int_3 = eq_1.int(); eq_1 = None | |
mul_6 = int_3 * 1000; int_3 = None | |
ne_1 = torch.ne(l_valid_past_index_, 0); l_valid_past_index_ = None | |
int_4 = ne_1.int(); ne_1 = None | |
mul_7 = int_4 * add_4; int_4 = None | |
add_5 = mul_6 + mul_7; mul_6 = mul_7 = None | |
return (add_1, getitem_1, add_4, add_5) | |
[2023-07-06 04:20:44,186] torch._dynamo.output_graph.__graph: [DEBUG] TRACED GRAPH | |
__compiled_fn_3 <eval_with_key>.111 opcode name target args kwargs | |
------------- ------------------------------- ------------------------------------------------------- -------------------------------------------------- ------------------------- | |
placeholder l_hidden_states_ L_hidden_states_ () {} | |
placeholder l_position_ids_ L_position_ids_ () {} | |
placeholder l_valid_past_index_ L_valid_past_index_ () {} | |
call_module l__self___qkv_proj L__self___qkv_proj (l_hidden_states_,) {} | |
call_method view view (l__self___qkv_proj, 1, 1000, 3, 4, 4) {} | |
call_method permute permute (view, 2, 0, 3, 1, 4) {} | |
call_function getitem <built-in function getitem> (permute, 0) {} | |
call_function getitem_1 <built-in function getitem> (permute, slice(1, None, None)) {} | |
call_function add <built-in function add> (l_valid_past_index_, 1000) {} | |
get_attr l__self___rotary_emb_cos_cached L__self___rotary_emb_cos_cached () {} | |
call_method to to (l__self___rotary_emb_cos_cached,) {'dtype': torch.float16} | |
get_attr l__self___rotary_emb_sin_cached L__self___rotary_emb_sin_cached () {} | |
call_method to_1 to (l__self___rotary_emb_sin_cached,) {'dtype': torch.float16} | |
call_function getitem_2 <built-in function getitem> (getitem_1, 0) {} | |
call_method squeeze squeeze (to, 1) {} | |
call_method squeeze_1 squeeze (squeeze, 0) {} | |
call_method squeeze_2 squeeze (to_1, 1) {} | |
call_method squeeze_3 squeeze (squeeze_2, 0) {} | |
call_function getitem_3 <built-in function getitem> (squeeze_1, l_position_ids_) {} | |
call_method unsqueeze unsqueeze (getitem_3, 1) {} | |
call_function getitem_4 <built-in function getitem> (squeeze_3, l_position_ids_) {} | |
call_method unsqueeze_1 unsqueeze (getitem_4, 1) {} | |
call_function mul <built-in function mul> (getitem, unsqueeze) {} | |
call_function roll <built-in method roll of type object at 0x7fb940469580> (getitem,) {'shifts': 2, 'dims': -1} | |
call_function getitem_5 <built-in function getitem> (roll, (Ellipsis, slice(None, 2, None))) {} | |
call_function imul <built-in function imul> (getitem_5, -1) {} | |
call_function setitem <built-in function setitem> (roll, (Ellipsis, slice(None, 2, None)), imul) {} | |
call_function mul_1 <built-in function mul> (roll, unsqueeze_1) {} | |
call_function add_1 <built-in function add> (mul, mul_1) {} | |
call_function mul_2 <built-in function mul> (getitem_2, unsqueeze) {} | |
call_function roll_1 <built-in method roll of type object at 0x7fb940469580> (getitem_2,) {'shifts': 2, 'dims': -1} | |
call_function getitem_6 <built-in function getitem> (roll_1, (Ellipsis, slice(None, 2, None))) {} | |
call_function imul_1 <built-in function imul> (getitem_6, -1) {} | |
call_function setitem_1 <built-in function setitem> (roll_1, (Ellipsis, slice(None, 2, None)), imul_1) {} | |
call_function mul_3 <built-in function mul> (roll_1, unsqueeze_1) {} | |
call_function add_2 <built-in function add> (mul_2, mul_3) {} | |
call_method copy_ copy_ (getitem_2, add_2) {} | |
call_function eq <built-in method eq of type object at 0x7fb940469580> (l_valid_past_index_, 0) {} | |
call_method int_1 int (eq,) {} | |
call_function mul_4 <built-in function mul> (int_1, 1000) {} | |
call_function ne <built-in method ne of type object at 0x7fb940469580> (l_valid_past_index_, 0) {} | |
call_method int_2 int (ne,) {} | |
call_function add_3 <built-in function add> (l_valid_past_index_, 1) {} | |
call_function mul_5 <built-in function mul> (int_2, add_3) {} | |
call_function add_4 <built-in function add> (mul_4, mul_5) {} | |
call_function eq_1 <built-in method eq of type object at 0x7fb940469580> (l_valid_past_index_, 0) {} | |
call_method int_3 int (eq_1,) {} | |
call_function mul_6 <built-in function mul> (int_3, 1000) {} | |
call_function ne_1 <built-in method ne of type object at 0x7fb940469580> (l_valid_past_index_, 0) {} | |
call_method int_4 int (ne_1,) {} | |
call_function mul_7 <built-in function mul> (int_4, add_4) {} | |
call_function add_5 <built-in function add> (mul_6, mul_7) {} | |
output output output ((add_1, getitem_1, add_4, add_5),) {} | |
[2023-07-06 04:20:44,188] torch._dynamo.output_graph: [INFO] Step 2: calling compiler function inductor | |
[2023-07-06 04:20:44,314] torch._functorch.aot_autograd.__aot_graphs: [INFO] TRACED GRAPH | |
===== Forward graph 14 ===== | |
<eval_with_key>.115 class <lambda>(torch.nn.Module): | |
def forward(self, arg0_1: f16[48, 16], arg1_1: f32[1, 1, 2048, 4], arg2_1: f32[1, 1, 2048, 4], arg3_1: f16[1, 1000, 16], arg4_1: i64[1, 1000], arg5_1: i64[]): | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:205, code: query_key_value_states = self.qkv_proj(hidden_states).view(bsz, q_len, 3, self.num_heads, self.head_dim).permute(2, 0, 3, 1, 4) | |
permute: f16[16, 48] = torch.ops.aten.permute.default(arg0_1, [1, 0]); arg0_1 = None | |
view: f16[1000, 16] = torch.ops.aten.view.default(arg3_1, [1000, 16]); arg3_1 = None | |
mm: f16[1000, 48] = torch.ops.aten.mm.default(view, permute); view = permute = None | |
view_1: f16[1, 1000, 48] = torch.ops.aten.view.default(mm, [1, 1000, 48]) | |
view_2: f16[1, 1000, 3, 4, 4] = torch.ops.aten.view.default(view_1, [1, 1000, 3, 4, 4]); view_1 = None | |
permute_1: f16[3, 1, 4, 1000, 4] = torch.ops.aten.permute.default(view_2, [2, 0, 3, 1, 4]); view_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:207, code: query_states = query_key_value_states[0] | |
select: f16[1, 4, 1000, 4] = torch.ops.aten.select.int(permute_1, 0, 0) | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:208, code: key_value_states = query_key_value_states[1:] | |
slice_1: f16[2, 1, 4, 1000, 4] = torch.ops.aten.slice.Tensor(permute_1, 0, 1, 9223372036854775807); permute_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:122, code: self.cos_cached.to(dtype=x.dtype), | |
convert_element_type: f16[1, 1, 2048, 4] = torch.ops.prims.convert_element_type.default(arg1_1, torch.float16); arg1_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:123, code: self.sin_cached.to(dtype=x.dtype), | |
convert_element_type_1: f16[1, 1, 2048, 4] = torch.ops.prims.convert_element_type.default(arg2_1, torch.float16); arg2_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:212, code: query_states = apply_rotary_pos_emb_opt(query_states, key_value_states[0], cos, sin, position_ids) | |
select_1: f16[1, 4, 1000, 4] = torch.ops.aten.select.int(slice_1, 0, 0); slice_1 = None | |
# No stacktrace found for following nodes | |
squeeze: f16[1, 2048, 4] = torch.ops.aten.squeeze.dim(convert_element_type, 1); convert_element_type = None | |
squeeze_1: f16[2048, 4] = torch.ops.aten.squeeze.dim(squeeze, 0); squeeze = None | |
squeeze_2: f16[1, 2048, 4] = torch.ops.aten.squeeze.dim(convert_element_type_1, 1); convert_element_type_1 = None | |
squeeze_3: f16[2048, 4] = torch.ops.aten.squeeze.dim(squeeze_2, 0); squeeze_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:141, code: cos = cos[position_ids].unsqueeze(1) # [bs, 1, seq_len, dim] | |
index: f16[1, 1000, 4] = torch.ops.aten.index.Tensor(squeeze_1, [arg4_1]); squeeze_1 = None | |
unsqueeze: f16[1, 1, 1000, 4] = torch.ops.aten.unsqueeze.default(index, 1); index = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:142, code: sin = sin[position_ids].unsqueeze(1) # [bs, 1, seq_len, dim] | |
index_1: f16[1, 1000, 4] = torch.ops.aten.index.Tensor(squeeze_3, [arg4_1]); squeeze_3 = arg4_1 = None | |
unsqueeze_1: f16[1, 1, 1000, 4] = torch.ops.aten.unsqueeze.default(index_1, 1); index_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:143, code: q_embed = (q * cos) + (rotate_half(q) * sin) | |
mul: f16[1, 4, 1000, 4] = torch.ops.aten.mul.Tensor(select, unsqueeze) | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:129, code: x = torch.roll(x, shifts=x.shape[-1] // 2, dims=-1) | |
roll: f16[1, 4, 1000, 4] = torch.ops.aten.roll.default(select, [2], [-1]); select = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:130, code: x[..., : x.shape[-1] // 2] *= -1 | |
slice_2: f16[1, 4, 1000, 2] = torch.ops.aten.slice.Tensor(roll, 3, 0, 2) | |
mul_1: f16[1, 4, 1000, 2] = torch.ops.aten.mul.Tensor(slice_2, -1); slice_2 = None | |
slice_scatter: f16[1, 4, 1000, 4] = torch.ops.aten.slice_scatter.default(roll, mul_1, 3, 0, 2); roll = mul_1 = None | |
slice_3: f16[1, 4, 1000, 2] = torch.ops.aten.slice.Tensor(slice_scatter, 3, 0, 2) | |
slice_scatter_1: f16[1, 4, 1000, 4] = torch.ops.aten.slice_scatter.default(slice_scatter, slice_3, 3, 0, 2); slice_scatter = slice_3 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:143, code: q_embed = (q * cos) + (rotate_half(q) * sin) | |
mul_2: f16[1, 4, 1000, 4] = torch.ops.aten.mul.Tensor(slice_scatter_1, unsqueeze_1); slice_scatter_1 = None | |
add_1: f16[1, 4, 1000, 4] = torch.ops.aten.add.Tensor(mul, mul_2); mul = mul_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:144, code: key_states.copy_((key_states * cos) + (rotate_half(key_states) * sin)) | |
mul_3: f16[1, 4, 1000, 4] = torch.ops.aten.mul.Tensor(select_1, unsqueeze); unsqueeze = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:129, code: x = torch.roll(x, shifts=x.shape[-1] // 2, dims=-1) | |
roll_1: f16[1, 4, 1000, 4] = torch.ops.aten.roll.default(select_1, [2], [-1]); select_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:130, code: x[..., : x.shape[-1] // 2] *= -1 | |
slice_7: f16[1, 4, 1000, 2] = torch.ops.aten.slice.Tensor(roll_1, 3, 0, 2) | |
mul_4: f16[1, 4, 1000, 2] = torch.ops.aten.mul.Tensor(slice_7, -1); slice_7 = None | |
slice_scatter_2: f16[1, 4, 1000, 4] = torch.ops.aten.slice_scatter.default(roll_1, mul_4, 3, 0, 2); roll_1 = mul_4 = None | |
slice_8: f16[1, 4, 1000, 2] = torch.ops.aten.slice.Tensor(slice_scatter_2, 3, 0, 2) | |
slice_scatter_3: f16[1, 4, 1000, 4] = torch.ops.aten.slice_scatter.default(slice_scatter_2, slice_8, 3, 0, 2); slice_scatter_2 = slice_8 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:144, code: key_states.copy_((key_states * cos) + (rotate_half(key_states) * sin)) | |
mul_5: f16[1, 4, 1000, 4] = torch.ops.aten.mul.Tensor(slice_scatter_3, unsqueeze_1); slice_scatter_3 = unsqueeze_1 = None | |
add_2: f16[1, 4, 1000, 4] = torch.ops.aten.add.Tensor(mul_3, mul_5); mul_3 = mul_5 = None | |
view_3: f16[1, 1000, 48] = torch.ops.aten.view.default(mm, [1, 1000, 48]); mm = None | |
view_4: f16[1, 1000, 3, 4, 4] = torch.ops.aten.view.default(view_3, [1, 1000, 3, 4, 4]); view_3 = None | |
permute_2: f16[3, 1, 4, 1000, 4] = torch.ops.aten.permute.default(view_4, [2, 0, 3, 1, 4]); view_4 = None | |
slice_12: f16[2, 1, 4, 1000, 4] = torch.ops.aten.slice.Tensor(permute_2, 0, 1, 9223372036854775807) | |
select_scatter: f16[2, 1, 4, 1000, 4] = torch.ops.aten.select_scatter.default(slice_12, add_2, 0, 0); slice_12 = add_2 = None | |
slice_scatter_4: f16[3, 1, 4, 1000, 4] = torch.ops.aten.slice_scatter.default(permute_2, select_scatter, 0, 1, 9223372036854775807); permute_2 = select_scatter = None | |
permute_3: f16[1, 1000, 3, 4, 4] = torch.ops.aten.permute.default(slice_scatter_4, [1, 3, 0, 2, 4]); slice_scatter_4 = None | |
view_5: f16[1, 1000, 48] = torch.ops.aten.view.default(permute_3, [1, 1000, 48]); permute_3 = None | |
view_6: f16[1000, 48] = torch.ops.aten.view.default(view_5, [1000, 48]); view_5 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:216, code: past_kv_slice_end = torch.eq(valid_past_index, 0).int() * q_len + torch.ne(valid_past_index, 0).int() * (valid_past_index + 1) | |
eq: b8[] = torch.ops.aten.eq.Scalar(arg5_1, 0) | |
convert_element_type_2: i32[] = torch.ops.prims.convert_element_type.default(eq, torch.int32); eq = None | |
mul_6: i32[] = torch.ops.aten.mul.Tensor(convert_element_type_2, 1000); convert_element_type_2 = None | |
ne: b8[] = torch.ops.aten.ne.Scalar(arg5_1, 0) | |
convert_element_type_3: i32[] = torch.ops.prims.convert_element_type.default(ne, torch.int32); ne = None | |
add_3: i64[] = torch.ops.aten.add.Tensor(arg5_1, 1) | |
mul_7: i64[] = torch.ops.aten.mul.Tensor(convert_element_type_3, add_3); convert_element_type_3 = add_3 = None | |
add_4: i64[] = torch.ops.aten.add.Tensor(mul_6, mul_7); mul_6 = mul_7 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:217, code: past_state_slice_end = torch.eq(valid_past_index, 0).int() * key_value_states.shape[-2] + torch.ne(valid_past_index, 0).int() * (past_kv_slice_end) | |
eq_1: b8[] = torch.ops.aten.eq.Scalar(arg5_1, 0) | |
convert_element_type_4: i32[] = torch.ops.prims.convert_element_type.default(eq_1, torch.int32); eq_1 = None | |
mul_8: i32[] = torch.ops.aten.mul.Tensor(convert_element_type_4, 1000); convert_element_type_4 = None | |
ne_1: b8[] = torch.ops.aten.ne.Scalar(arg5_1, 0); arg5_1 = None | |
convert_element_type_5: i32[] = torch.ops.prims.convert_element_type.default(ne_1, torch.int32); ne_1 = None | |
mul_9: i64[] = torch.ops.aten.mul.Tensor(convert_element_type_5, add_4); convert_element_type_5 = None | |
add_5: i64[] = torch.ops.aten.add.Tensor(mul_8, mul_9); mul_8 = mul_9 = None | |
# No stacktrace found for following nodes | |
view_9: f16[1, 1000, 48] = torch.ops.aten.view.default(view_6, [1, 1000, 48]); view_6 = None | |
view_10: f16[1, 1000, 3, 4, 4] = torch.ops.aten.view.default(view_9, [1, 1000, 3, 4, 4]); view_9 = None | |
permute_5: f16[3, 1, 4, 1000, 4] = torch.ops.aten.permute.default(view_10, [2, 0, 3, 1, 4]); view_10 = None | |
slice_14: f16[2, 1, 4, 1000, 4] = torch.ops.aten.slice.Tensor(permute_5, 0, 1, 9223372036854775807); permute_5 = None | |
return (add_1, slice_14, add_4, add_5) | |
[2023-07-06 04:20:44,343] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 1 | |
[2023-07-06 04:20:44,505] torch._inductor.ir: [DEBUG] convert_to_reinterpret_view failed: stride=[16000, 0, 4000, 4, 1] offset=16000 index=16000*r0 + 4000*r1 + 4*r2 + r3 + 16000*(((4*r1 + r3)//16)) + 16000 | |
[2023-07-06 04:20:44,648] torch._inductor.graph: [DEBUG] Force channels last inputs for 0 conv for the current graph with id 1 | |
[2023-07-06 04:20:49,653] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 2 configs | |
[2023-07-06 04:20:49,653] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 512, num_warps: 4, num_stages: 1 | |
[2023-07-06 04:20:49,653] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 256, num_warps: 4, num_stages: 1 | |
[2023-07-06 04:20:49,653] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 2 configs | |
[2023-07-06 04:20:49,653] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 256, num_warps: 4, num_stages: 1 | |
[2023-07-06 04:20:49,653] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 128, num_warps: 4, num_stages: 1 | |
[2023-07-06 04:20:49,655] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 2 configs | |
[2023-07-06 04:20:49,655] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 256, num_warps: 4, num_stages: 1 | |
[2023-07-06 04:20:49,655] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 128, num_warps: 4, num_stages: 1 | |
[2023-07-06 04:20:50,220] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 2 configs | |
[2023-07-06 04:20:50,220] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 256, num_warps: 4, num_stages: 1 | |
[2023-07-06 04:20:50,220] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 128, num_warps: 4, num_stages: 1 | |
[2023-07-06 04:20:50,774] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 2 configs | |
[2023-07-06 04:20:50,774] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 512, num_warps: 4, num_stages: 1 | |
[2023-07-06 04:20:50,774] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 256, num_warps: 4, num_stages: 1 | |
[2023-07-06 04:20:50,785] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 2 configs | |
[2023-07-06 04:20:50,785] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 256, num_warps: 4, num_stages: 1 | |
[2023-07-06 04:20:50,785] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 128, num_warps: 4, num_stages: 1 | |
[2023-07-06 04:20:50,793] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_ubuntu/lo/clozylkw544bb7yqlyamdizaia7jo2uesdfqh4qhrfm2er5ahkkn.py | |
[2023-07-06 04:20:50,794] torch._inductor.compile_fx: [DEBUG] disabled cudagraphs because ['non-cuda device in graph'] | |
skipping cudagraphs due to multiple devices | |
[2023-07-06 04:20:50,794] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 1 | |
[2023-07-06 04:20:50,796] torch._dynamo.output_graph: [INFO] Step 2: done compiler function inductor | |
[2023-07-06 04:20:50,806] torch._dynamo.convert_frame.__guards: [DEBUG] GUARDS: | |
___check_obj_id(L['self'], 140431662763168) | |
hasattr(L['position_ids'], '_dynamo_dynamic_indices') == False | |
hasattr(L['hidden_states'], '_dynamo_dynamic_indices') == False | |
hasattr(L['attention_mask'], '_dynamo_dynamic_indices') == False | |
hasattr(L['past_key_value'], '_dynamo_dynamic_indices') == False | |
hasattr(L['valid_past_index'], '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['output_attentions'], 7698112) | |
not ___is_grad_enabled() | |
not ___are_deterministic_algorithms_enabled() | |
utils_device.CURRENT_DEVICE == None | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks.keys()) == set() | |
[2023-07-06 04:20:51,067] torch._inductor.triton_heuristics: [DEBUG] Benchmark all input configs get: | |
[2023-07-06 04:20:51,067] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 256, num_warps: 4, num_stages: 1: 0.006144, nreg 56, nspill 0, #shared-mem 512 | |
[2023-07-06 04:20:51,067] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 128, num_warps: 4, num_stages: 1: 0.005120, nreg 32, nspill 0, #shared-mem 0 | |
[2023-07-06 04:20:51,067] torch._inductor.triton_heuristics: [DEBUG] Save heuristic tuning result to /tmp/torchinductor_ubuntu/bg/cbgmas5pmuel253gz3ahwbawktqf6aqc24qu3fx2cnoj45poporq.best_config | |
[2023-07-06 04:20:51,154] torch._inductor.triton_heuristics: [DEBUG] Benchmark all input configs get: | |
[2023-07-06 04:20:51,155] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 512, num_warps: 4, num_stages: 1: 0.005120, nreg 20, nspill 0, #shared-mem 0 | |
[2023-07-06 04:20:51,155] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 256, num_warps: 4, num_stages: 1: 0.005120, nreg 14, nspill 0, #shared-mem 0 | |
[2023-07-06 04:20:51,155] torch._inductor.triton_heuristics: [DEBUG] Save heuristic tuning result to /tmp/torchinductor_ubuntu/5p/c5px4pduotxc7ykett376w6ivfllsxhle4zfciermbzwclsz3whs.best_config | |
[2023-07-06 04:20:51,242] torch._inductor.triton_heuristics: [DEBUG] Benchmark all input configs get: | |
[2023-07-06 04:20:51,242] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 256, num_warps: 4, num_stages: 1: 0.005120, nreg 17, nspill 0, #shared-mem 0 | |
[2023-07-06 04:20:51,242] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 128, num_warps: 4, num_stages: 1: 0.005120, nreg 13, nspill 0, #shared-mem 0 | |
[2023-07-06 04:20:51,242] torch._inductor.triton_heuristics: [DEBUG] Save heuristic tuning result to /tmp/torchinductor_ubuntu/tu/ctum23fv2ehp4y2k4x7ymm3ssw7lch2csweiyhwgxheu4g67rwic.best_config | |
[2023-07-06 04:20:51,242] torch._dynamo.eval_frame: [DEBUG] skipping __iter__ /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_tensor.py | |
[2023-07-06 04:20:51,279] torch._dynamo.eval_frame: [DEBUG] skipping forward /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/linear.py | |
[2023-07-06 04:20:51,280] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo start tracing <resume in forward> /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:286 | |
[2023-07-06 04:20:51,280] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['___stack0'][0] (1, 1000, 16) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None] | |
[2023-07-06 04:20:51,281] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['residual'] (1, 1000, 16) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None] | |
[2023-07-06 04:20:51,282] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST ___stack0 [] | |
[2023-07-06 04:20:51,282] torch._dynamo.symbolic_convert: [DEBUG] TRACE JUMP_ABSOLUTE 38 [TupleVariable()] | |
[2023-07-06 04:20:51,282] torch._dynamo.symbolic_convert: [DEBUG] TRACE UNPACK_SEQUENCE 3 [TupleVariable()] | |
[2023-07-06 04:20:51,282] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [ConstantVariable(NoneType), ConstantVariable(NoneType), TensorVariable()] | |
[2023-07-06 04:20:51,282] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST self_attn_weights [ConstantVariable(NoneType), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:51,282] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST present_key_value [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:51,282] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:294 | |
[2023-07-06 04:20:51,282] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST residual [] | |
[2023-07-06 04:20:51,282] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:51,282] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:51,283] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:51,284] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:297 | |
[2023-07-06 04:20:51,284] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:51,284] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST residual [TensorVariable()] | |
[2023-07-06 04:20:51,284] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:298 | |
[2023-07-06 04:20:51,284] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:51,284] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR post_attention_layernorm [NNModuleVariable()] | |
[2023-07-06 04:20:51,284] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [NNModuleVariable()] | |
[2023-07-06 04:20:51,284] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:51,285] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:51,286] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:51,286] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:51,286] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:51,287] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:51,287] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:51,287] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:51,287] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:51,287] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:51,287] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable())] | |
[2023-07-06 04:20:51,287] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:51,287] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:51,288] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:51,288] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:51,288] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:51,288] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:51,288] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:51,288] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:51,288] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:51,289] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:51,289] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:51,289] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:51,289] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:51,289] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:51,289] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:51,289] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:51,289] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:51,289] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:51,289] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:51,289] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:51,290] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:51,290] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:51,290] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:51,290] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:51,290] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:51,290] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:51,290] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:51,290] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:51,290] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:51,290] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:51,290] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:51,290] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable())] | |
[2023-07-06 04:20:51,290] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:51,290] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:51,290] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:51,290] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:51,291] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object forward at 0x7fb8cddd7ec0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 83> | |
[2023-07-06 04:20:51,291] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:84 | |
[2023-07-06 04:20:51,291] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:51,291] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR dtype [TensorVariable()] | |
[2023-07-06 04:20:51,291] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST input_dtype [TorchVariable(torch.float16)] | |
[2023-07-06 04:20:51,291] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85 | |
[2023-07-06 04:20:51,291] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:51,291] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable()] | |
[2023-07-06 04:20:51,292] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:51,292] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR float32 [GetAttrVariable(TensorVariable(), to), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:51,292] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), to), ConstantVariable(dtype)] | |
[2023-07-06 04:20:51,293] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR pow [TensorVariable()] | |
[2023-07-06 04:20:51,293] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [GetAttrVariable(TensorVariable(), pow)] | |
[2023-07-06 04:20:51,293] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), pow), ConstantVariable(int)] | |
[2023-07-06 04:20:51,294] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR mean [TensorVariable()] | |
[2023-07-06 04:20:51,295] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [GetAttrVariable(TensorVariable(), mean)] | |
[2023-07-06 04:20:51,295] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST True [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int)] | |
[2023-07-06 04:20:51,295] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('keepdim',) [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int), ConstantVariable(bool)] | |
[2023-07-06 04:20:51,295] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 2 [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int), ConstantVariable(bool), ConstantVariable(tuple)] | |
[2023-07-06 04:20:51,296] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST variance [TensorVariable()] | |
[2023-07-06 04:20:51,296] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86 | |
[2023-07-06 04:20:51,296] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:51,296] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [TensorVariable()] | |
[2023-07-06 04:20:51,296] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR rsqrt [TensorVariable(), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:51,296] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST variance [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:51,296] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:51,296] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR variance_epsilon [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:51,297] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(float)] | |
[2023-07-06 04:20:51,297] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:51,298] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:51,299] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:51,299] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88 | |
[2023-07-06 04:20:51,299] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:51,299] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR weight [NNModuleVariable()] | |
[2023-07-06 04:20:51,299] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['self'].post_attention_layernorm.weight (16,) [<DimDynamic.STATIC: 2>] [None] | |
[2023-07-06 04:20:51,300] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:51,300] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:51,301] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable()] | |
[2023-07-06 04:20:51,301] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST input_dtype [GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:51,301] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16)] | |
[2023-07-06 04:20:51,302] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:51,302] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object forward at 0x7fb8cddd7ec0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 83> | |
[2023-07-06 04:20:51,302] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:51,302] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:51,302] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:51,302] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:299 | |
[2023-07-06 04:20:51,302] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:51,302] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR mlp [NNModuleVariable()] | |
[2023-07-06 04:20:51,303] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [NNModuleVariable()] | |
[2023-07-06 04:20:51,303] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:51,303] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:51,305] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:51,305] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:51,305] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:51,306] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:51,306] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:51,306] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:51,306] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:51,306] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:51,307] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function LlamaMLP.forward at 0x7fb8d0c464d0>, NNModuleVariable())] | |
[2023-07-06 04:20:51,307] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:51,307] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:51,307] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:51,307] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:51,307] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:51,307] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:51,307] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:51,307] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:51,307] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:51,308] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:51,308] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:51,308] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:51,308] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:51,308] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:51,308] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:51,308] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:51,308] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:51,308] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:51,308] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:51,308] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:51,308] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:51,308] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:51,308] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:51,309] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:51,309] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:51,309] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:51,309] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:51,309] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:51,309] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:51,309] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:51,309] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:51,309] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function LlamaMLP.forward at 0x7fb8d0c464d0>, NNModuleVariable())] | |
[2023-07-06 04:20:51,309] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function LlamaMLP.forward at 0x7fb8d0c464d0>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:51,309] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function LlamaMLP.forward at 0x7fb8d0c464d0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:51,309] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function LlamaMLP.forward at 0x7fb8d0c464d0>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:51,309] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function LlamaMLP.forward at 0x7fb8d0c464d0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:51,309] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object forward at 0x7fb8cddf85b0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 161> | |
[2023-07-06 04:20:51,309] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:162 | |
[2023-07-06 04:20:51,310] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:51,310] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR down_proj [NNModuleVariable()] | |
[2023-07-06 04:20:51,310] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [NNModuleVariable()] | |
[2023-07-06 04:20:51,310] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR act_fn [NNModuleVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:51,310] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [NNModuleVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:51,310] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR gate_proj [NNModuleVariable(), NNModuleVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:51,311] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [NNModuleVariable(), NNModuleVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:51,311] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), NNModuleVariable(), NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:51,314] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:51,314] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:51,315] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:51,315] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:51,315] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:51,316] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:51,316] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:51,316] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:51,316] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:51,316] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:51,316] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function SiLUActivation.forward at 0x7fb8ccf64280>, NNModuleVariable())] | |
[2023-07-06 04:20:51,316] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:51,316] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:51,316] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:51,317] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:51,317] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:51,317] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:51,317] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:51,317] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:51,317] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:51,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:51,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:51,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:51,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:51,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:51,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:51,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:51,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:51,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:51,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:51,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:51,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:51,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:51,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:51,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:51,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:51,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:51,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:51,318] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:51,319] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:51,319] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:51,319] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:51,319] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function SiLUActivation.forward at 0x7fb8ccf64280>, NNModuleVariable())] | |
[2023-07-06 04:20:51,319] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function SiLUActivation.forward at 0x7fb8ccf64280>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:51,319] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function SiLUActivation.forward at 0x7fb8ccf64280>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:51,319] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function SiLUActivation.forward at 0x7fb8ccf64280>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:51,319] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function SiLUActivation.forward at 0x7fb8ccf64280>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:51,319] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object forward at 0x7fb8ccf62760, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/activations.py", line 149> | |
[2023-07-06 04:20:51,319] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/activations.py:150 | |
[2023-07-06 04:20:51,319] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL nn [] | |
[2023-07-06 04:20:51,320] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR functional [TorchVariable(<module 'torch.nn' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/__init__.py'>)] | |
[2023-07-06 04:20:51,321] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR silu [TorchVariable(<module 'torch.nn.functional' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/functional.py'>)] | |
[2023-07-06 04:20:51,321] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST input [TorchVariable(<function silu at 0x7fb945bc5990>)] | |
[2023-07-06 04:20:51,321] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [TorchVariable(<function silu at 0x7fb945bc5990>), TensorVariable()] | |
[2023-07-06 04:20:51,325] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:51,325] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object forward at 0x7fb8ccf62760, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/activations.py", line 149> | |
[2023-07-06 04:20:51,325] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:51,325] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:51,325] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:51,325] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR up_proj [NNModuleVariable(), TensorVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:51,326] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [NNModuleVariable(), TensorVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:51,326] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable(), NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:51,328] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [NNModuleVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:51,329] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:51,332] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:51,332] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object forward at 0x7fb8cddf85b0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 161> | |
[2023-07-06 04:20:51,332] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:51,332] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:51,332] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:51,332] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:300 | |
[2023-07-06 04:20:51,332] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST residual [] | |
[2023-07-06 04:20:51,332] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:51,332] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:51,333] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:51,333] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:302 | |
[2023-07-06 04:20:51,333] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:51,333] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_TUPLE 1 [TensorVariable()] | |
[2023-07-06 04:20:51,333] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST outputs [TupleVariable()] | |
[2023-07-06 04:20:51,333] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:304 | |
[2023-07-06 04:20:51,333] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST output_attentions [] | |
[2023-07-06 04:20:51,333] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 106 [ConstantVariable(bool)] | |
[2023-07-06 04:20:51,333] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:307 | |
[2023-07-06 04:20:51,333] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST outputs [] | |
[2023-07-06 04:20:51,333] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST present_key_value [TupleVariable()] | |
[2023-07-06 04:20:51,333] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_TUPLE 1 [TupleVariable(), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:51,333] torch._dynamo.symbolic_convert: [DEBUG] TRACE INPLACE_ADD None [TupleVariable(), TupleVariable()] | |
[2023-07-06 04:20:51,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST outputs [TupleVariable()] | |
[2023-07-06 04:20:51,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:309 | |
[2023-07-06 04:20:51,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST outputs [] | |
[2023-07-06 04:20:51,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TupleVariable()] | |
[2023-07-06 04:20:51,334] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo done tracing <resume in forward> (RETURN_VALUE) | |
[2023-07-06 04:20:51,334] torch._dynamo.symbolic_convert: [DEBUG] RETURN_VALUE triggered compile | |
[2023-07-06 04:20:51,334] torch._dynamo.output_graph: [DEBUG] COMPILING GRAPH due to GraphCompileReason(reason='return_value', user_stack=[<FrameSummary file /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py, line 309 in <resume in forward>>], graph_break=False) | |
[2023-07-06 04:20:51,336] torch._dynamo.output_graph.__graph_code: [DEBUG] TRACED GRAPH | |
===== __compiled_fn_4 ===== | |
<eval_with_key>.120 class GraphModule(torch.nn.Module): | |
def forward(self, L_stack0_0_ : torch.Tensor, L_residual_ : torch.Tensor): | |
l_stack0_0_ = L_stack0_0_ | |
l_residual_ = L_residual_ | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:294, code: hidden_states = residual + hidden_states | |
add = l_residual_ + l_stack0_0_; l_residual_ = l_stack0_0_ = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85, code: variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) | |
to = add.to(torch.float32) | |
pow_1 = to.pow(2); to = None | |
mean = pow_1.mean(-1, keepdim = True); pow_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86, code: hidden_states = hidden_states * torch.rsqrt(variance + self.variance_epsilon) | |
add_1 = mean + 1e-06; mean = None | |
rsqrt = torch.rsqrt(add_1); add_1 = None | |
mul = add * rsqrt; rsqrt = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88, code: return (self.weight * hidden_states).to(input_dtype) | |
l__self___post_attention_layernorm_weight = self.L__self___post_attention_layernorm_weight | |
mul_1 = l__self___post_attention_layernorm_weight * mul; l__self___post_attention_layernorm_weight = mul = None | |
to_1 = mul_1.to(torch.float16); mul_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:162, code: return self.down_proj(self.act_fn(self.gate_proj(x)) * self.up_proj(x)) | |
l__self___mlp_gate_proj = self.L__self___mlp_gate_proj(to_1) | |
# File: /home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/activations.py:150, code: return nn.functional.silu(input) | |
silu = torch.nn.functional.silu(l__self___mlp_gate_proj); l__self___mlp_gate_proj = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:162, code: return self.down_proj(self.act_fn(self.gate_proj(x)) * self.up_proj(x)) | |
l__self___mlp_up_proj = self.L__self___mlp_up_proj(to_1); to_1 = None | |
mul_2 = silu * l__self___mlp_up_proj; silu = l__self___mlp_up_proj = None | |
l__self___mlp_down_proj = self.L__self___mlp_down_proj(mul_2); mul_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:300, code: hidden_states = residual + hidden_states | |
add_2 = add + l__self___mlp_down_proj; add = l__self___mlp_down_proj = None | |
return (add_2,) | |
[2023-07-06 04:20:51,336] torch._dynamo.output_graph.__graph: [DEBUG] TRACED GRAPH | |
__compiled_fn_4 <eval_with_key>.120 opcode name target args kwargs | |
------------- ----------------------------------------- -------------------------------------------------------- ------------------------------------------------ ----------------- | |
placeholder l_stack0_0_ L_stack0_0_ () {} | |
placeholder l_residual_ L_residual_ () {} | |
call_function add <built-in function add> (l_residual_, l_stack0_0_) {} | |
call_method to to (add, torch.float32) {} | |
call_method pow_1 pow (to, 2) {} | |
call_method mean mean (pow_1, -1) {'keepdim': True} | |
call_function add_1 <built-in function add> (mean, 1e-06) {} | |
call_function rsqrt <built-in method rsqrt of type object at 0x7fb940469580> (add_1,) {} | |
call_function mul <built-in function mul> (add, rsqrt) {} | |
get_attr l__self___post_attention_layernorm_weight L__self___post_attention_layernorm_weight () {} | |
call_function mul_1 <built-in function mul> (l__self___post_attention_layernorm_weight, mul) {} | |
call_method to_1 to (mul_1, torch.float16) {} | |
call_module l__self___mlp_gate_proj L__self___mlp_gate_proj (to_1,) {} | |
call_function silu <function silu at 0x7fb945bc5990> (l__self___mlp_gate_proj,) {} | |
call_module l__self___mlp_up_proj L__self___mlp_up_proj (to_1,) {} | |
call_function mul_2 <built-in function mul> (silu, l__self___mlp_up_proj) {} | |
call_module l__self___mlp_down_proj L__self___mlp_down_proj (mul_2,) {} | |
call_function add_2 <built-in function add> (add, l__self___mlp_down_proj) {} | |
output output output ((add_2,),) {} | |
[2023-07-06 04:20:51,337] torch._dynamo.output_graph: [INFO] Step 2: calling compiler function inductor | |
[2023-07-06 04:20:51,385] torch._functorch.aot_autograd.__aot_graphs: [INFO] TRACED GRAPH | |
===== Forward graph 15 ===== | |
<eval_with_key>.124 class <lambda>(torch.nn.Module): | |
def forward(self, arg0_1: f16[16], arg1_1: f16[64, 16], arg2_1: f16[64, 16], arg3_1: f16[16, 64], arg4_1: f16[1, 1000, 16], arg5_1: f16[1, 1000, 16]): | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:294, code: hidden_states = residual + hidden_states | |
add: f16[1, 1000, 16] = torch.ops.aten.add.Tensor(arg5_1, arg4_1); arg5_1 = arg4_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85, code: variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) | |
convert_element_type: f32[1, 1000, 16] = torch.ops.prims.convert_element_type.default(add, torch.float32) | |
pow_1: f32[1, 1000, 16] = torch.ops.aten.pow.Tensor_Scalar(convert_element_type, 2); convert_element_type = None | |
mean: f32[1, 1000, 1] = torch.ops.aten.mean.dim(pow_1, [-1], True); pow_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86, code: hidden_states = hidden_states * torch.rsqrt(variance + self.variance_epsilon) | |
add_1: f32[1, 1000, 1] = torch.ops.aten.add.Tensor(mean, 1e-06); mean = None | |
rsqrt: f32[1, 1000, 1] = torch.ops.aten.rsqrt.default(add_1); add_1 = None | |
mul: f32[1, 1000, 16] = torch.ops.aten.mul.Tensor(add, rsqrt); rsqrt = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88, code: return (self.weight * hidden_states).to(input_dtype) | |
mul_1: f32[1, 1000, 16] = torch.ops.aten.mul.Tensor(arg0_1, mul); arg0_1 = mul = None | |
convert_element_type_1: f16[1, 1000, 16] = torch.ops.prims.convert_element_type.default(mul_1, torch.float16); mul_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:162, code: return self.down_proj(self.act_fn(self.gate_proj(x)) * self.up_proj(x)) | |
permute: f16[16, 64] = torch.ops.aten.permute.default(arg1_1, [1, 0]); arg1_1 = None | |
view: f16[1000, 16] = torch.ops.aten.view.default(convert_element_type_1, [1000, 16]) | |
mm: f16[1000, 64] = torch.ops.aten.mm.default(view, permute); view = permute = None | |
view_1: f16[1, 1000, 64] = torch.ops.aten.view.default(mm, [1, 1000, 64]); mm = None | |
# File: /home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/activations.py:150, code: return nn.functional.silu(input) | |
convert_element_type_2: f32[1, 1000, 64] = torch.ops.prims.convert_element_type.default(view_1, torch.float32); view_1 = None | |
sigmoid: f32[1, 1000, 64] = torch.ops.aten.sigmoid.default(convert_element_type_2) | |
mul_2: f32[1, 1000, 64] = torch.ops.aten.mul.Tensor(convert_element_type_2, sigmoid); convert_element_type_2 = sigmoid = None | |
convert_element_type_3: f16[1, 1000, 64] = torch.ops.prims.convert_element_type.default(mul_2, torch.float16); mul_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:162, code: return self.down_proj(self.act_fn(self.gate_proj(x)) * self.up_proj(x)) | |
permute_1: f16[16, 64] = torch.ops.aten.permute.default(arg2_1, [1, 0]); arg2_1 = None | |
view_2: f16[1000, 16] = torch.ops.aten.view.default(convert_element_type_1, [1000, 16]); convert_element_type_1 = None | |
mm_1: f16[1000, 64] = torch.ops.aten.mm.default(view_2, permute_1); view_2 = permute_1 = None | |
view_3: f16[1, 1000, 64] = torch.ops.aten.view.default(mm_1, [1, 1000, 64]); mm_1 = None | |
mul_3: f16[1, 1000, 64] = torch.ops.aten.mul.Tensor(convert_element_type_3, view_3); convert_element_type_3 = view_3 = None | |
permute_2: f16[64, 16] = torch.ops.aten.permute.default(arg3_1, [1, 0]); arg3_1 = None | |
view_4: f16[1000, 64] = torch.ops.aten.view.default(mul_3, [1000, 64]); mul_3 = None | |
mm_2: f16[1000, 16] = torch.ops.aten.mm.default(view_4, permute_2); view_4 = permute_2 = None | |
view_5: f16[1, 1000, 16] = torch.ops.aten.view.default(mm_2, [1, 1000, 16]); mm_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:300, code: hidden_states = residual + hidden_states | |
add_2: f16[1, 1000, 16] = torch.ops.aten.add.Tensor(add, view_5); add = view_5 = None | |
return (add_2,) | |
[2023-07-06 04:20:51,413] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 2 | |
[2023-07-06 04:20:51,459] torch._inductor.graph: [DEBUG] Force channels last inputs for 0 conv for the current graph with id 2 | |
[2023-07-06 04:20:51,486] torch._inductor.scheduler: [DEBUG] remove_buffer('buf0') | |
[2023-07-06 04:20:52,116] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 2 configs | |
[2023-07-06 04:20:52,116] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 256, num_warps: 4, num_stages: 1 | |
[2023-07-06 04:20:52,116] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 128, num_warps: 4, num_stages: 1 | |
[2023-07-06 04:20:52,124] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 2 configs | |
[2023-07-06 04:20:52,125] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 512, num_warps: 4, num_stages: 1 | |
[2023-07-06 04:20:52,125] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 256, num_warps: 4, num_stages: 1 | |
[2023-07-06 04:20:52,125] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 4 configs | |
[2023-07-06 04:20:52,125] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 1, num_warps: 2, num_stages: 1 | |
[2023-07-06 04:20:52,125] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 8, num_warps: 2, num_stages: 1 | |
[2023-07-06 04:20:52,125] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 32, num_warps: 4, num_stages: 1 | |
[2023-07-06 04:20:52,125] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 128, num_warps: 8, num_stages: 1 | |
[2023-07-06 04:20:52,988] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 4 configs | |
[2023-07-06 04:20:52,988] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 1, num_warps: 2, num_stages: 1 | |
[2023-07-06 04:20:52,988] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 8, num_warps: 2, num_stages: 1 | |
[2023-07-06 04:20:52,988] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 32, num_warps: 4, num_stages: 1 | |
[2023-07-06 04:20:52,988] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 128, num_warps: 8, num_stages: 1 | |
[2023-07-06 04:20:53,009] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 2 configs | |
[2023-07-06 04:20:53,009] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 512, num_warps: 4, num_stages: 1 | |
[2023-07-06 04:20:53,009] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 256, num_warps: 4, num_stages: 1 | |
[2023-07-06 04:20:53,018] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 2 configs | |
[2023-07-06 04:20:53,018] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 256, num_warps: 4, num_stages: 1 | |
[2023-07-06 04:20:53,019] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 128, num_warps: 4, num_stages: 1 | |
[2023-07-06 04:20:53,026] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_ubuntu/or/cor2fodulphywxrtc7zlvp6tsjseyifwlaiv6ticnqodficz2elj.py | |
[2023-07-06 04:20:53,026] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 2 | |
[2023-07-06 04:20:53,029] torch._dynamo.output_graph: [INFO] Step 2: done compiler function inductor | |
[2023-07-06 04:20:53,037] torch._dynamo.convert_frame.__guards: [DEBUG] GUARDS: | |
___check_obj_id(L['self'], 140431662763072) | |
hasattr(L['residual'], '_dynamo_dynamic_indices') == False | |
___check_type_id(L['___stack0'], 7638272) | |
len(L['___stack0']) == 3 | |
hasattr(L['___stack0'][0], '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['___stack0'][1], 7649088) | |
___check_obj_id(L['___stack0'][2], 7649088) | |
___check_obj_id(L['output_attentions'], 7698112) | |
not ___is_grad_enabled() | |
not ___are_deterministic_algorithms_enabled() | |
utils_device.CURRENT_DEVICE == None | |
str(G['torch'].float32) == 'torch.float32' | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks.keys()) == set() | |
[2023-07-06 04:20:53,037] torch._inductor.cudagraph_trees: [DEBUG] Running warmup of function 1 | |
[2023-07-06 04:20:53,213] torch._inductor.triton_heuristics: [DEBUG] Benchmark all input configs get: | |
[2023-07-06 04:20:53,213] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 1, num_warps: 2, num_stages: 1: 0.005120, nreg 18, nspill 0, #shared-mem 256 | |
[2023-07-06 04:20:53,213] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 8, num_warps: 2, num_stages: 1: 0.005120, nreg 18, nspill 0, #shared-mem 256 | |
[2023-07-06 04:20:53,213] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 32, num_warps: 4, num_stages: 1: 0.005120, nreg 22, nspill 0, #shared-mem 512 | |
[2023-07-06 04:20:53,213] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 128, num_warps: 8, num_stages: 1: 0.005120, nreg 28, nspill 0, #shared-mem 1024 | |
[2023-07-06 04:20:53,214] torch._inductor.triton_heuristics: [DEBUG] Save heuristic tuning result to /tmp/torchinductor_ubuntu/eh/ceh7jbqrlrzcv3zqh4rrhwapliqluprt3lrh53lvvuhdpwvqyn7m.best_config | |
[2023-07-06 04:20:53,305] torch._inductor.triton_heuristics: [DEBUG] Benchmark all input configs get: | |
[2023-07-06 04:20:53,305] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 512, num_warps: 4, num_stages: 1: 0.009216, nreg 19, nspill 0, #shared-mem 0 | |
[2023-07-06 04:20:53,305] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 256, num_warps: 4, num_stages: 1: 0.009216, nreg 16, nspill 0, #shared-mem 0 | |
[2023-07-06 04:20:53,305] torch._inductor.triton_heuristics: [DEBUG] Save heuristic tuning result to /tmp/torchinductor_ubuntu/v5/cv57f2eqjbxmcec2s7yyljqzlj4o3t34kuus254ilu5x7nwh4o4o.best_config | |
[2023-07-06 04:20:53,397] torch._inductor.triton_heuristics: [DEBUG] Benchmark all input configs get: | |
[2023-07-06 04:20:53,397] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 256, num_warps: 4, num_stages: 1: 0.007168, nreg 14, nspill 0, #shared-mem 0 | |
[2023-07-06 04:20:53,397] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 128, num_warps: 4, num_stages: 1: 0.007184, nreg 13, nspill 0, #shared-mem 0 | |
[2023-07-06 04:20:53,398] torch._inductor.triton_heuristics: [DEBUG] Save heuristic tuning result to /tmp/torchinductor_ubuntu/yq/cyqg2c2iwzykrwd2ftiynpnxqx2veaxu4jqvchxyg2wxm6kg37hz.best_config | |
[2023-07-06 04:20:53,398] torch._dynamo.eval_frame: [DEBUG] skipping __call__ /home/ubuntu/miniconda3/lib/python3.10/weakref.py | |
[2023-07-06 04:20:53,398] torch._dynamo.eval_frame: [DEBUG] skipping del_ten /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_subclasses/meta_utils.py | |
[2023-07-06 04:20:53,398] torch._dynamo.eval_frame: [DEBUG] skipping pop /home/ubuntu/miniconda3/lib/python3.10/weakref.py | |
[2023-07-06 04:20:53,398] torch._dynamo.eval_frame: [DEBUG] skipping __hash__ /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/utils/weak.py | |
[2023-07-06 04:20:53,398] torch._dynamo.eval_frame: [DEBUG] skipping expired /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/multiprocessing/reductions.py | |
[2023-07-06 04:20:53,398] torch._dynamo.eval_frame: [DEBUG] skipping _expired /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/storage.py | |
[2023-07-06 04:20:53,398] torch._dynamo.convert_frame.__recompiles: [DEBUG] ('Recompiling function forward in /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py', 'set env var TORCHDYNAMO_REPORT_GUARD_FAILURES=1 to debug further') | |
[2023-07-06 04:20:53,398] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo start tracing forward /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:261 | |
[2023-07-06 04:20:53,399] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['hidden_states'] (1, 1000, 16) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None] | |
[2023-07-06 04:20:53,401] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['attention_mask'] (1, 1003) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None] | |
[2023-07-06 04:20:53,401] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['position_ids'] (1, 1000) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None] | |
[2023-07-06 04:20:53,402] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['past_key_value'] (2, 1, 4, 1003, 4) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None, None, None] | |
[2023-07-06 04:20:53,403] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['valid_past_index'] () [] [] | |
[2023-07-06 04:20:53,404] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:281 | |
[2023-07-06 04:20:53,404] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:53,404] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST residual [TensorVariable()] | |
[2023-07-06 04:20:53,404] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:283 | |
[2023-07-06 04:20:53,404] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:53,404] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR input_layernorm [NNModuleVariable()] | |
[2023-07-06 04:20:53,404] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [NNModuleVariable()] | |
[2023-07-06 04:20:53,404] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,405] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:53,406] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:53,406] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:53,406] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:53,406] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:53,407] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:53,407] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:53,407] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:53,407] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:53,407] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable())] | |
[2023-07-06 04:20:53,407] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:53,407] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:53,407] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:53,408] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:53,408] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:53,408] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:53,408] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:53,408] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:53,408] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:53,408] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:53,408] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:53,408] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:53,409] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:53,409] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:53,409] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:53,409] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:53,409] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:53,409] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:53,409] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:53,409] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:53,409] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:53,409] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:53,409] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:53,409] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:53,409] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:53,409] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:53,409] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:53,410] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:53,410] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:53,410] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:53,410] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:53,410] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable())] | |
[2023-07-06 04:20:53,410] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:53,410] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:53,410] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:53,410] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:53,410] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object forward at 0x7fb8cddd7ec0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 83> | |
[2023-07-06 04:20:53,411] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:84 | |
[2023-07-06 04:20:53,411] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:53,411] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR dtype [TensorVariable()] | |
[2023-07-06 04:20:53,413] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST input_dtype [TorchVariable(torch.float16)] | |
[2023-07-06 04:20:53,413] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85 | |
[2023-07-06 04:20:53,413] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:53,413] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable()] | |
[2023-07-06 04:20:53,414] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:53,414] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR float32 [GetAttrVariable(TensorVariable(), to), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:53,414] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), to), ConstantVariable(dtype)] | |
[2023-07-06 04:20:53,415] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR pow [TensorVariable()] | |
[2023-07-06 04:20:53,416] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [GetAttrVariable(TensorVariable(), pow)] | |
[2023-07-06 04:20:53,416] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), pow), ConstantVariable(int)] | |
[2023-07-06 04:20:53,417] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR mean [TensorVariable()] | |
[2023-07-06 04:20:53,417] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [GetAttrVariable(TensorVariable(), mean)] | |
[2023-07-06 04:20:53,417] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST True [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int)] | |
[2023-07-06 04:20:53,417] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('keepdim',) [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int), ConstantVariable(bool)] | |
[2023-07-06 04:20:53,417] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 2 [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int), ConstantVariable(bool), ConstantVariable(tuple)] | |
[2023-07-06 04:20:53,418] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST variance [TensorVariable()] | |
[2023-07-06 04:20:53,418] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86 | |
[2023-07-06 04:20:53,418] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:53,418] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [TensorVariable()] | |
[2023-07-06 04:20:53,418] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR rsqrt [TensorVariable(), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:53,418] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST variance [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:53,419] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:53,419] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR variance_epsilon [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:53,419] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(float)] | |
[2023-07-06 04:20:53,420] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:53,420] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,421] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:53,421] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88 | |
[2023-07-06 04:20:53,421] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:53,421] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR weight [NNModuleVariable()] | |
[2023-07-06 04:20:53,422] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['self'].input_layernorm.weight (16,) [<DimDynamic.STATIC: 2>] [None] | |
[2023-07-06 04:20:53,422] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:53,422] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,423] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable()] | |
[2023-07-06 04:20:53,423] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST input_dtype [GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:53,423] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16)] | |
[2023-07-06 04:20:53,424] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:53,425] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object forward at 0x7fb8cddd7ec0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 83> | |
[2023-07-06 04:20:53,425] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:53,425] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:53,425] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:53,425] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:286 | |
[2023-07-06 04:20:53,425] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:53,425] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR self_attn [NNModuleVariable()] | |
[2023-07-06 04:20:53,425] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:287 | |
[2023-07-06 04:20:53,425] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [NNModuleVariable()] | |
[2023-07-06 04:20:53,425] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:288 | |
[2023-07-06 04:20:53,425] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST attention_mask [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,425] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:289 | |
[2023-07-06 04:20:53,425] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST position_ids [NNModuleVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,425] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:290 | |
[2023-07-06 04:20:53,425] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST past_key_value [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,425] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:291 | |
[2023-07-06 04:20:53,425] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST output_attentions [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,426] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:292 | |
[2023-07-06 04:20:53,426] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ConstantVariable(bool)] | |
[2023-07-06 04:20:53,426] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:286 | |
[2023-07-06 04:20:53,426] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('hidden_states', 'attention_mask', 'position_ids', 'past_key_value', 'output_attentions', 'valid_past_index') [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ConstantVariable(bool), TensorVariable()] | |
[2023-07-06 04:20:53,426] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 6 [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ConstantVariable(bool), TensorVariable(), ConstantVariable(tuple)] | |
[2023-07-06 04:20:53,426] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:53,427] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:53,427] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:53,428] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:53,428] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:53,428] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:53,428] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:53,428] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:53,428] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:53,429] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function LlamaAttention.forward at 0x7fb8d0c46680>, NNModuleVariable())] | |
[2023-07-06 04:20:53,429] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:53,429] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:53,429] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:53,429] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:53,429] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:53,429] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:53,429] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:53,429] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:53,429] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:53,430] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:53,430] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:53,430] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:53,430] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:53,430] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:53,430] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:53,430] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:53,430] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:53,430] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:53,430] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:53,430] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:53,430] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:53,430] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:53,430] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:53,430] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:53,431] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:53,431] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:53,431] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:53,431] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:53,431] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:53,431] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:53,431] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:53,431] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function LlamaAttention.forward at 0x7fb8d0c46680>, NNModuleVariable())] | |
[2023-07-06 04:20:53,431] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function LlamaAttention.forward at 0x7fb8d0c46680>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:53,431] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function LlamaAttention.forward at 0x7fb8d0c46680>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:53,431] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function LlamaAttention.forward at 0x7fb8d0c46680>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:53,431] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function LlamaAttention.forward at 0x7fb8d0c46680>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:53,431] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:53,432] torch._dynamo.symbolic_convert: [DEBUG] FAILED INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:53,432] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:53,432] torch._dynamo.symbolic_convert: [DEBUG] break_graph_if_unsupported triggered compile | |
Traceback (most recent call last): | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2187, in inline_call_ | |
sub_locals, closure_cells = func.bind_args(parent, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 187, in bind_args | |
[ | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 188, in <listcomp> | |
wrap(val=arg, source=source) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 62, in wrap_bound_arg | |
return VariableBuilder(tx, source=source, **options)(val) | |
TypeError: VariableBuilder.__init__() got an unexpected keyword argument 'guards' | |
During handling of the above exception, another exception occurred: | |
Traceback (most recent call last): | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 390, in wrapper | |
return inner_fn(self, inst) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 1152, in CALL_FUNCTION_KW | |
self.call_function(fn, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 559, in call_function | |
self.push(fn.call_function(self, args, kwargs)) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/nn_module.py", line 333, in call_function | |
return tx.inline_user_function_return( | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 595, in inline_user_function_return | |
result = InliningInstructionTranslator.inline_call(self, fn, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2132, in inline_call | |
return cls.inline_call_(parent, func, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2229, in inline_call_ | |
tracer.run() | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 708, in run | |
and self.step() | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 668, in step | |
getattr(self, inst.opname)(inst) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 390, in wrapper | |
return inner_fn(self, inst) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 1140, in CALL_FUNCTION_EX | |
self.call_function(fn, argsvars.items, kwargsvars.items) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 559, in call_function | |
self.push(fn.call_function(self, args, kwargs)) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 333, in call_function | |
return super().call_function(tx, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 287, in call_function | |
return super().call_function(tx, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 120, in call_function | |
return tx.inline_user_function_return( | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 595, in inline_user_function_return | |
result = InliningInstructionTranslator.inline_call(self, fn, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2132, in inline_call | |
return cls.inline_call_(parent, func, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2190, in inline_call_ | |
raise ArgsMismatchError( | |
torch._dynamo.exc.ArgsMismatchError: VariableBuilder.__init__() got an unexpected keyword argument 'guards'. | |
func = 'forward' /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:190, args = [<class 'trfs_fast.llama.LlamaAttention'>], kwargs = OrderedDict([('hidden_states', TensorVariable()), ('attention_mask', TensorVariable()), ('position_ids', TensorVariable()), ('past_key_value', TensorVariable()), ('output_attentions', ConstantVariable(bool)), ('valid_past_index', TensorVariable())]) | |
[2023-07-06 04:20:53,432] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:53,432] torch._dynamo.output_graph: [DEBUG] COMPILING GRAPH due to GraphCompileReason(reason="VariableBuilder.__init__() got an unexpected keyword argument 'guards'.\n func = 'forward' /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:190, args = [<class 'trfs_fast.llama.LlamaAttention'>], kwargs = OrderedDict([('hidden_states', TensorVariable()), ('attention_mask', TensorVariable()), ('position_ids', TensorVariable()), ('past_key_value', TensorVariable()), ('output_attentions', ConstantVariable(bool)), ('valid_past_index', TensorVariable())])", user_stack=[<FrameSummary file /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py, line 286 in forward>, <FrameSummary file /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py, line 1514 in _call_impl>], graph_break=True) | |
[2023-07-06 04:20:53,433] torch._dynamo.output_graph: [DEBUG] REMOVE UNUSED GRAPHARG L['attention_mask'] | |
[2023-07-06 04:20:53,433] torch._dynamo.output_graph: [DEBUG] REMOVE UNUSED GRAPHARG L['position_ids'] | |
[2023-07-06 04:20:53,433] torch._dynamo.output_graph: [DEBUG] REMOVE UNUSED GRAPHARG L['past_key_value'] | |
[2023-07-06 04:20:53,434] torch._dynamo.output_graph: [DEBUG] REMOVE UNUSED GRAPHARG L['valid_past_index'] | |
[2023-07-06 04:20:53,434] torch._dynamo.output_graph.__graph_code: [DEBUG] TRACED GRAPH | |
===== __compiled_fn_5 ===== | |
<eval_with_key>.133 class GraphModule(torch.nn.Module): | |
def forward(self, L_hidden_states_ : torch.Tensor): | |
l_hidden_states_ = L_hidden_states_ | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85, code: variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) | |
to = l_hidden_states_.to(torch.float32) | |
pow_1 = to.pow(2); to = None | |
mean = pow_1.mean(-1, keepdim = True); pow_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86, code: hidden_states = hidden_states * torch.rsqrt(variance + self.variance_epsilon) | |
add = mean + 1e-06; mean = None | |
rsqrt = torch.rsqrt(add); add = None | |
mul = l_hidden_states_ * rsqrt; l_hidden_states_ = rsqrt = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88, code: return (self.weight * hidden_states).to(input_dtype) | |
l__self___input_layernorm_weight = self.L__self___input_layernorm_weight | |
mul_1 = l__self___input_layernorm_weight * mul; l__self___input_layernorm_weight = mul = None | |
to_1 = mul_1.to(torch.float16); mul_1 = None | |
return (to_1,) | |
[2023-07-06 04:20:53,434] torch._dynamo.output_graph.__graph: [DEBUG] TRACED GRAPH | |
__compiled_fn_5 <eval_with_key>.133 opcode name target args kwargs | |
------------- -------------------------------- -------------------------------------------------------- --------------------------------------- ----------------- | |
placeholder l_hidden_states_ L_hidden_states_ () {} | |
call_method to to (l_hidden_states_, torch.float32) {} | |
call_method pow_1 pow (to, 2) {} | |
call_method mean mean (pow_1, -1) {'keepdim': True} | |
call_function add <built-in function add> (mean, 1e-06) {} | |
call_function rsqrt <built-in method rsqrt of type object at 0x7fb940469580> (add,) {} | |
call_function mul <built-in function mul> (l_hidden_states_, rsqrt) {} | |
get_attr l__self___input_layernorm_weight L__self___input_layernorm_weight () {} | |
call_function mul_1 <built-in function mul> (l__self___input_layernorm_weight, mul) {} | |
call_method to_1 to (mul_1, torch.float16) {} | |
output output output ((to_1,),) {} | |
[2023-07-06 04:20:53,435] torch._dynamo.output_graph: [INFO] Step 2: calling compiler function inductor | |
[2023-07-06 04:20:53,454] torch._functorch.aot_autograd.__aot_graphs: [INFO] TRACED GRAPH | |
===== Forward graph 16 ===== | |
<eval_with_key>.137 class <lambda>(torch.nn.Module): | |
def forward(self, arg0_1: f16[16], arg1_1: f16[1, 1000, 16]): | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85, code: variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) | |
convert_element_type: f32[1, 1000, 16] = torch.ops.prims.convert_element_type.default(arg1_1, torch.float32) | |
pow_1: f32[1, 1000, 16] = torch.ops.aten.pow.Tensor_Scalar(convert_element_type, 2); convert_element_type = None | |
mean: f32[1, 1000, 1] = torch.ops.aten.mean.dim(pow_1, [-1], True); pow_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86, code: hidden_states = hidden_states * torch.rsqrt(variance + self.variance_epsilon) | |
add: f32[1, 1000, 1] = torch.ops.aten.add.Tensor(mean, 1e-06); mean = None | |
rsqrt: f32[1, 1000, 1] = torch.ops.aten.rsqrt.default(add); add = None | |
mul: f32[1, 1000, 16] = torch.ops.aten.mul.Tensor(arg1_1, rsqrt); arg1_1 = rsqrt = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88, code: return (self.weight * hidden_states).to(input_dtype) | |
mul_1: f32[1, 1000, 16] = torch.ops.aten.mul.Tensor(arg0_1, mul); arg0_1 = mul = None | |
convert_element_type_1: f16[1, 1000, 16] = torch.ops.prims.convert_element_type.default(mul_1, torch.float16); mul_1 = None | |
return (convert_element_type_1,) | |
[2023-07-06 04:20:53,455] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 3 | |
[2023-07-06 04:20:53,472] torch._inductor.graph: [DEBUG] Force channels last inputs for 0 conv for the current graph with id 3 | |
[2023-07-06 04:20:53,487] torch._inductor.scheduler: [DEBUG] remove_buffer('buf0') | |
[2023-07-06 04:20:53,489] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_ubuntu/tn/ctnjyv37bz6fvtmwelga5764xq272722mvbuxkb4qqgd3syt7hwl.py | |
[2023-07-06 04:20:53,489] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 3 | |
[2023-07-06 04:20:53,491] torch._dynamo.output_graph: [INFO] Step 2: done compiler function inductor | |
[2023-07-06 04:20:53,499] torch._dynamo.convert_frame.__guards: [DEBUG] GUARDS: | |
___check_obj_id(L['self'], 140431662486464) | |
hasattr(L['position_ids'], '_dynamo_dynamic_indices') == False | |
___check_type_id(L['hidden_states'], 77765680) | |
hasattr(L['hidden_states'], '_dynamo_dynamic_indices') == False | |
hasattr(L['attention_mask'], '_dynamo_dynamic_indices') == False | |
hasattr(L['past_key_value'], '_dynamo_dynamic_indices') == False | |
hasattr(L['valid_past_index'], '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['output_attentions'], 7698112) | |
not ___is_grad_enabled() | |
not ___are_deterministic_algorithms_enabled() | |
utils_device.CURRENT_DEVICE == None | |
str(G['torch'].float32) == 'torch.float32' | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks.keys()) == set() | |
[2023-07-06 04:20:53,499] torch._inductor.cudagraph_trees: [DEBUG] Running warmup of function 2 | |
[2023-07-06 04:20:53,500] torch._dynamo.convert_frame.__recompiles: [DEBUG] ('Recompiling function forward in /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py', 'set env var TORCHDYNAMO_REPORT_GUARD_FAILURES=1 to debug further') | |
[2023-07-06 04:20:53,501] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo start tracing forward /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:190 | |
[2023-07-06 04:20:53,502] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['hidden_states'] (1, 1000, 16) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None] | |
[2023-07-06 04:20:53,503] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['attention_mask'] (1, 1003) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None] | |
[2023-07-06 04:20:53,503] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['position_ids'] (1, 1000) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None] | |
[2023-07-06 04:20:53,504] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['past_key_value'] (2, 1, 4, 1003, 4) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None, None, None] | |
[2023-07-06 04:20:53,505] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['valid_past_index'] () [] [] | |
[2023-07-06 04:20:53,506] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:199 | |
[2023-07-06 04:20:53,506] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST output_attentions [] | |
[2023-07-06 04:20:53,506] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST True [ConstantVariable(bool)] | |
[2023-07-06 04:20:53,506] torch._dynamo.symbolic_convert: [DEBUG] TRACE IS_OP 0 [ConstantVariable(bool), ConstantVariable(bool)] | |
[2023-07-06 04:20:53,506] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(bool)] | |
[2023-07-06 04:20:53,506] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:202 | |
[2023-07-06 04:20:53,506] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:53,506] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR size [TensorVariable()] | |
[2023-07-06 04:20:53,506] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [GetAttrVariable(TensorVariable(), size)] | |
[2023-07-06 04:20:53,507] torch._dynamo.symbolic_convert: [DEBUG] TRACE UNPACK_SEQUENCE 3 [SizeVariable()] | |
[2023-07-06 04:20:53,507] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST bsz [ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:53,507] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST q_len [ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:53,507] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST _ [ConstantVariable(int)] | |
[2023-07-06 04:20:53,507] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:205 | |
[2023-07-06 04:20:53,507] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:53,507] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR qkv_proj [NNModuleVariable()] | |
[2023-07-06 04:20:53,507] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [NNModuleVariable()] | |
[2023-07-06 04:20:53,507] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,510] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR view [TensorVariable()] | |
[2023-07-06 04:20:53,510] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST bsz [GetAttrVariable(TensorVariable(), view)] | |
[2023-07-06 04:20:53,510] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST q_len [GetAttrVariable(TensorVariable(), view), ConstantVariable(int)] | |
[2023-07-06 04:20:53,510] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 3 [GetAttrVariable(TensorVariable(), view), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:53,510] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [GetAttrVariable(TensorVariable(), view), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:53,510] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR num_heads [GetAttrVariable(TensorVariable(), view), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), NNModuleVariable()] | |
[2023-07-06 04:20:53,511] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [GetAttrVariable(TensorVariable(), view), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:53,511] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR head_dim [GetAttrVariable(TensorVariable(), view), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), NNModuleVariable()] | |
[2023-07-06 04:20:53,511] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 5 [GetAttrVariable(TensorVariable(), view), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:53,512] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR permute [TensorVariable()] | |
[2023-07-06 04:20:53,512] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [GetAttrVariable(TensorVariable(), permute)] | |
[2023-07-06 04:20:53,512] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [GetAttrVariable(TensorVariable(), permute), ConstantVariable(int)] | |
[2023-07-06 04:20:53,512] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 3 [GetAttrVariable(TensorVariable(), permute), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:53,512] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [GetAttrVariable(TensorVariable(), permute), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:53,512] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 4 [GetAttrVariable(TensorVariable(), permute), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:53,512] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 5 [GetAttrVariable(TensorVariable(), permute), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:53,513] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST query_key_value_states [TensorVariable()] | |
[2023-07-06 04:20:53,513] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:207 | |
[2023-07-06 04:20:53,513] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST query_key_value_states [] | |
[2023-07-06 04:20:53,513] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [TensorVariable()] | |
[2023-07-06 04:20:53,513] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:53,514] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST query_states [TensorVariable()] | |
[2023-07-06 04:20:53,514] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:208 | |
[2023-07-06 04:20:53,514] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST query_key_value_states [] | |
[2023-07-06 04:20:53,514] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [TensorVariable()] | |
[2023-07-06 04:20:53,514] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:53,514] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_SLICE 2 [TensorVariable(), ConstantVariable(int), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:53,514] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), SliceVariable()] | |
[2023-07-06 04:20:53,515] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST key_value_states [TensorVariable()] | |
[2023-07-06 04:20:53,515] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:211 | |
[2023-07-06 04:20:53,516] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:53,516] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR rotary_emb [NNModuleVariable()] | |
[2023-07-06 04:20:53,516] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_value_states [NNModuleVariable()] | |
[2023-07-06 04:20:53,516] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,516] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST q_len [NNModuleVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,516] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [NNModuleVariable(), TensorVariable(), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:53,517] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('seq_len',) [NNModuleVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,517] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 2 [NNModuleVariable(), TensorVariable(), TensorVariable(), ConstantVariable(tuple)] | |
[2023-07-06 04:20:53,517] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:53,519] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:53,519] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:53,519] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:53,519] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:53,519] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:53,520] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:53,520] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:53,520] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:53,520] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function LlamaRotaryEmbedding.forward at 0x7fb8d0c46290>, NNModuleVariable())] | |
[2023-07-06 04:20:53,520] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:53,520] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:53,520] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:53,520] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:53,520] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:53,520] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:53,521] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:53,521] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:53,521] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:53,521] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:53,521] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:53,521] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:53,522] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:53,522] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:53,522] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:53,522] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:53,522] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:53,522] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:53,522] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:53,522] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:53,522] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:53,522] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:53,522] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:53,522] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:53,522] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:53,522] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:53,522] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:53,522] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:53,522] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:53,522] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:53,523] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:53,523] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function LlamaRotaryEmbedding.forward at 0x7fb8d0c46290>, NNModuleVariable())] | |
[2023-07-06 04:20:53,523] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function LlamaRotaryEmbedding.forward at 0x7fb8d0c46290>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:53,523] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function LlamaRotaryEmbedding.forward at 0x7fb8d0c46290>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:53,523] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function LlamaRotaryEmbedding.forward at 0x7fb8d0c46290>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:53,523] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function LlamaRotaryEmbedding.forward at 0x7fb8d0c46290>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:53,523] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object forward at 0x7fb8cddf80e0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 107> | |
[2023-07-06 04:20:53,523] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:122 | |
[2023-07-06 04:20:53,523] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:53,523] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR cos_cached [NNModuleVariable()] | |
[2023-07-06 04:20:53,524] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['self'].rotary_emb.cos_cached (1, 1, 2048, 4) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None, None] | |
[2023-07-06 04:20:53,524] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable()] | |
[2023-07-06 04:20:53,525] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:53,525] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR dtype [GetAttrVariable(TensorVariable(), to), TensorVariable()] | |
[2023-07-06 04:20:53,525] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('dtype',) [GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16)] | |
[2023-07-06 04:20:53,525] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 1 [GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16), ConstantVariable(tuple)] | |
[2023-07-06 04:20:53,526] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:123 | |
[2023-07-06 04:20:53,526] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [TensorVariable()] | |
[2023-07-06 04:20:53,526] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR sin_cached [TensorVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:53,527] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['self'].rotary_emb.sin_cached (1, 1, 2048, 4) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None, None] | |
[2023-07-06 04:20:53,527] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,527] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TensorVariable(), GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:53,527] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR dtype [TensorVariable(), GetAttrVariable(TensorVariable(), to), TensorVariable()] | |
[2023-07-06 04:20:53,528] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('dtype',) [TensorVariable(), GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16)] | |
[2023-07-06 04:20:53,528] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 1 [TensorVariable(), GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16), ConstantVariable(tuple)] | |
[2023-07-06 04:20:53,529] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:121 | |
[2023-07-06 04:20:53,529] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_TUPLE 2 [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,529] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TupleVariable()] | |
[2023-07-06 04:20:53,529] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object forward at 0x7fb8cddf80e0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 107> | |
[2023-07-06 04:20:53,529] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TupleVariable()] | |
[2023-07-06 04:20:53,529] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:53,529] torch._dynamo.symbolic_convert: [DEBUG] TRACE UNPACK_SEQUENCE 2 [TupleVariable()] | |
[2023-07-06 04:20:53,529] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST cos [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,529] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST sin [TensorVariable()] | |
[2023-07-06 04:20:53,529] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:212 | |
[2023-07-06 04:20:53,529] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL apply_rotary_pos_emb_opt [] | |
[2023-07-06 04:20:53,529] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST query_states [UserFunctionVariable()] | |
[2023-07-06 04:20:53,530] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_value_states [UserFunctionVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,530] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [UserFunctionVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,530] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [UserFunctionVariable(), TensorVariable(), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:53,530] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST cos [UserFunctionVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,530] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST sin [UserFunctionVariable(), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,530] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST position_ids [UserFunctionVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,530] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 5 [UserFunctionVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,531] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object apply_rotary_pos_emb_opt at 0x7fb8cddf82f0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 135> | |
[2023-07-06 04:20:53,531] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:139 | |
[2023-07-06 04:20:53,531] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST cos [] | |
[2023-07-06 04:20:53,531] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR squeeze [TensorVariable()] | |
[2023-07-06 04:20:53,531] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [GetAttrVariable(TensorVariable(), squeeze)] | |
[2023-07-06 04:20:53,531] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), squeeze), ConstantVariable(int)] | |
[2023-07-06 04:20:53,532] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR squeeze [TensorVariable()] | |
[2023-07-06 04:20:53,532] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [GetAttrVariable(TensorVariable(), squeeze)] | |
[2023-07-06 04:20:53,532] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), squeeze), ConstantVariable(int)] | |
[2023-07-06 04:20:53,533] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST cos [TensorVariable()] | |
[2023-07-06 04:20:53,533] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:140 | |
[2023-07-06 04:20:53,533] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST sin [] | |
[2023-07-06 04:20:53,533] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR squeeze [TensorVariable()] | |
[2023-07-06 04:20:53,533] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [GetAttrVariable(TensorVariable(), squeeze)] | |
[2023-07-06 04:20:53,533] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), squeeze), ConstantVariable(int)] | |
[2023-07-06 04:20:53,534] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR squeeze [TensorVariable()] | |
[2023-07-06 04:20:53,534] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [GetAttrVariable(TensorVariable(), squeeze)] | |
[2023-07-06 04:20:53,534] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), squeeze), ConstantVariable(int)] | |
[2023-07-06 04:20:53,535] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST sin [TensorVariable()] | |
[2023-07-06 04:20:53,535] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:141 | |
[2023-07-06 04:20:53,535] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST cos [] | |
[2023-07-06 04:20:53,535] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST position_ids [TensorVariable()] | |
[2023-07-06 04:20:53,535] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,536] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR unsqueeze [TensorVariable()] | |
[2023-07-06 04:20:53,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [GetAttrVariable(TensorVariable(), unsqueeze)] | |
[2023-07-06 04:20:53,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), unsqueeze), ConstantVariable(int)] | |
[2023-07-06 04:20:53,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST cos [TensorVariable()] | |
[2023-07-06 04:20:53,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:142 | |
[2023-07-06 04:20:53,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST sin [] | |
[2023-07-06 04:20:53,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST position_ids [TensorVariable()] | |
[2023-07-06 04:20:53,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,539] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR unsqueeze [TensorVariable()] | |
[2023-07-06 04:20:53,539] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [GetAttrVariable(TensorVariable(), unsqueeze)] | |
[2023-07-06 04:20:53,539] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), unsqueeze), ConstantVariable(int)] | |
[2023-07-06 04:20:53,540] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST sin [TensorVariable()] | |
[2023-07-06 04:20:53,540] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:143 | |
[2023-07-06 04:20:53,540] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST q [] | |
[2023-07-06 04:20:53,540] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST cos [TensorVariable()] | |
[2023-07-06 04:20:53,540] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,541] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL rotate_half [TensorVariable()] | |
[2023-07-06 04:20:53,541] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST q [TensorVariable(), UserFunctionVariable()] | |
[2023-07-06 04:20:53,541] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [TensorVariable(), UserFunctionVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,541] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object rotate_half at 0x7fb8cddf8240, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 127> | |
[2023-07-06 04:20:53,541] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:129 | |
[2023-07-06 04:20:53,541] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:53,542] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR roll [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:53,542] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TorchVariable(<built-in method roll of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:53,542] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:53,542] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR shape [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,542] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ShapeVariable()] | |
[2023-07-06 04:20:53,542] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ShapeVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:53,543] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:53,543] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_FLOOR_DIVIDE None [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:53,543] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:53,543] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('shifts', 'dims') [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:53,543] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 3 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int), ConstantVariable(int), ConstantVariable(tuple)] | |
[2023-07-06 04:20:53,544] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST x [TensorVariable()] | |
[2023-07-06 04:20:53,544] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:130 | |
[2023-07-06 04:20:53,544] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [] | |
[2023-07-06 04:20:53,544] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST Ellipsis [TensorVariable()] | |
[2023-07-06 04:20:53,544] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [TensorVariable(), ConstantVariable(ellipsis)] | |
[2023-07-06 04:20:53,544] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:53,544] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR shape [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), TensorVariable()] | |
[2023-07-06 04:20:53,545] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ShapeVariable()] | |
[2023-07-06 04:20:53,545] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ShapeVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:53,545] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ConstantVariable(int)] | |
[2023-07-06 04:20:53,545] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_FLOOR_DIVIDE None [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:53,545] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_SLICE 2 [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ConstantVariable(int)] | |
[2023-07-06 04:20:53,545] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_TUPLE 2 [TensorVariable(), ConstantVariable(ellipsis), SliceVariable()] | |
[2023-07-06 04:20:53,545] torch._dynamo.symbolic_convert: [DEBUG] TRACE DUP_TOP_TWO None [TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:53,545] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), TupleVariable(), TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:53,546] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TensorVariable(), TupleVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,546] torch._dynamo.symbolic_convert: [DEBUG] TRACE INPLACE_MULTIPLY None [TensorVariable(), TupleVariable(), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:53,547] torch._dynamo.symbolic_convert: [DEBUG] TRACE ROT_THREE None [TensorVariable(), TupleVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,547] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_SUBSCR None [TensorVariable(), TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:53,547] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:131 | |
[2023-07-06 04:20:53,547] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [] | |
[2023-07-06 04:20:53,547] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:53,547] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object rotate_half at 0x7fb8cddf8240, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 127> | |
[2023-07-06 04:20:53,547] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST sin [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,547] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,548] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,549] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST q_embed [TensorVariable()] | |
[2023-07-06 04:20:53,549] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:144 | |
[2023-07-06 04:20:53,549] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_states [] | |
[2023-07-06 04:20:53,549] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR copy_ [TensorVariable()] | |
[2023-07-06 04:20:53,549] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_states [GetAttrVariable(TensorVariable(), copy_)] | |
[2023-07-06 04:20:53,549] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST cos [GetAttrVariable(TensorVariable(), copy_), TensorVariable()] | |
[2023-07-06 04:20:53,549] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [GetAttrVariable(TensorVariable(), copy_), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL rotate_half [GetAttrVariable(TensorVariable(), copy_), TensorVariable()] | |
[2023-07-06 04:20:53,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_states [GetAttrVariable(TensorVariable(), copy_), TensorVariable(), UserFunctionVariable()] | |
[2023-07-06 04:20:53,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), copy_), TensorVariable(), UserFunctionVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,551] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object rotate_half at 0x7fb8cddf8240, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 127> | |
[2023-07-06 04:20:53,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:129 | |
[2023-07-06 04:20:53,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:53,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR roll [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:53,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TorchVariable(<built-in method roll of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:53,552] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:53,552] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR shape [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,552] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ShapeVariable()] | |
[2023-07-06 04:20:53,552] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ShapeVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:53,552] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:53,552] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_FLOOR_DIVIDE None [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:53,552] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:53,552] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('shifts', 'dims') [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:53,552] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 3 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int), ConstantVariable(int), ConstantVariable(tuple)] | |
[2023-07-06 04:20:53,553] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST x [TensorVariable()] | |
[2023-07-06 04:20:53,553] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:130 | |
[2023-07-06 04:20:53,553] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [] | |
[2023-07-06 04:20:53,554] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST Ellipsis [TensorVariable()] | |
[2023-07-06 04:20:53,554] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [TensorVariable(), ConstantVariable(ellipsis)] | |
[2023-07-06 04:20:53,554] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:53,554] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR shape [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), TensorVariable()] | |
[2023-07-06 04:20:53,554] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ShapeVariable()] | |
[2023-07-06 04:20:53,554] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ShapeVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:53,554] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ConstantVariable(int)] | |
[2023-07-06 04:20:53,554] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_FLOOR_DIVIDE None [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:53,554] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_SLICE 2 [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ConstantVariable(int)] | |
[2023-07-06 04:20:53,555] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_TUPLE 2 [TensorVariable(), ConstantVariable(ellipsis), SliceVariable()] | |
[2023-07-06 04:20:53,555] torch._dynamo.symbolic_convert: [DEBUG] TRACE DUP_TOP_TWO None [TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:53,555] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), TupleVariable(), TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:53,556] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TensorVariable(), TupleVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,556] torch._dynamo.symbolic_convert: [DEBUG] TRACE INPLACE_MULTIPLY None [TensorVariable(), TupleVariable(), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:53,556] torch._dynamo.symbolic_convert: [DEBUG] TRACE ROT_THREE None [TensorVariable(), TupleVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,556] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_SUBSCR None [TensorVariable(), TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:53,557] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:131 | |
[2023-07-06 04:20:53,557] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [] | |
[2023-07-06 04:20:53,557] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:53,557] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object rotate_half at 0x7fb8cddf8240, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 127> | |
[2023-07-06 04:20:53,557] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST sin [GetAttrVariable(TensorVariable(), copy_), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,557] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [GetAttrVariable(TensorVariable(), copy_), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,558] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [GetAttrVariable(TensorVariable(), copy_), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,559] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), copy_), TensorVariable()] | |
[2023-07-06 04:20:53,559] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_TOP None [TensorVariable()] | |
[2023-07-06 04:20:53,559] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:145 | |
[2023-07-06 04:20:53,559] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST q_embed [] | |
[2023-07-06 04:20:53,559] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:53,559] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object apply_rotary_pos_emb_opt at 0x7fb8cddf82f0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 135> | |
[2023-07-06 04:20:53,559] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST query_states [TensorVariable()] | |
[2023-07-06 04:20:53,559] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:215 | |
[2023-07-06 04:20:53,559] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [] | |
[2023-07-06 04:20:53,560] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST past_kv_slice_start [TensorVariable()] | |
[2023-07-06 04:20:53,560] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:216 | |
[2023-07-06 04:20:53,560] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:53,560] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR eq [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:53,560] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [TorchVariable(<built-in method eq of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:53,560] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [TorchVariable(<built-in method eq of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:53,560] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 2 [TorchVariable(<built-in method eq of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:53,561] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR int [TensorVariable()] | |
[2023-07-06 04:20:53,561] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [GetAttrVariable(TensorVariable(), int)] | |
[2023-07-06 04:20:53,562] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST q_len [TensorVariable()] | |
[2023-07-06 04:20:53,562] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:53,563] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [TensorVariable()] | |
[2023-07-06 04:20:53,563] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR ne [TensorVariable(), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:53,563] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [TensorVariable(), TorchVariable(<built-in method ne of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:53,563] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [TensorVariable(), TorchVariable(<built-in method ne of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:53,563] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 2 [TensorVariable(), TorchVariable(<built-in method ne of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:53,564] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR int [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,564] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TensorVariable(), GetAttrVariable(TensorVariable(), int)] | |
[2023-07-06 04:20:53,565] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,565] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,565] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TensorVariable(), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:53,566] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,567] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,568] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST past_kv_slice_end [TensorVariable()] | |
[2023-07-06 04:20:53,568] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:217 | |
[2023-07-06 04:20:53,568] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:53,568] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR eq [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:53,568] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [TorchVariable(<built-in method eq of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:53,568] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [TorchVariable(<built-in method eq of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:53,568] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 2 [TorchVariable(<built-in method eq of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:53,569] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR int [TensorVariable()] | |
[2023-07-06 04:20:53,569] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [GetAttrVariable(TensorVariable(), int)] | |
[2023-07-06 04:20:53,570] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_value_states [TensorVariable()] | |
[2023-07-06 04:20:53,571] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR shape [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,571] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -2 [TensorVariable(), ShapeVariable()] | |
[2023-07-06 04:20:53,571] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), ShapeVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:53,571] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:53,572] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [TensorVariable()] | |
[2023-07-06 04:20:53,572] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR ne [TensorVariable(), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:53,572] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [TensorVariable(), TorchVariable(<built-in method ne of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:53,572] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [TensorVariable(), TorchVariable(<built-in method ne of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:53,572] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 2 [TensorVariable(), TorchVariable(<built-in method ne of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:53,573] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR int [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,573] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TensorVariable(), GetAttrVariable(TensorVariable(), int)] | |
[2023-07-06 04:20:53,574] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST past_kv_slice_end [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,574] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,575] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST past_state_slice_end [TensorVariable()] | |
[2023-07-06 04:20:53,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:218 | |
[2023-07-06 04:20:53,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_value_states [] | |
[2023-07-06 04:20:53,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST past_key_value [TensorVariable()] | |
[2023-07-06 04:20:53,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST Ellipsis [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST past_kv_slice_start [TensorVariable(), TensorVariable(), ConstantVariable(ellipsis)] | |
[2023-07-06 04:20:53,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST past_kv_slice_end [TensorVariable(), TensorVariable(), ConstantVariable(ellipsis), TensorVariable()] | |
[2023-07-06 04:20:53,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_SLICE 2 [TensorVariable(), TensorVariable(), ConstantVariable(ellipsis), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:53,576] torch._dynamo.symbolic_convert: [DEBUG] step triggered compile | |
Traceback (most recent call last): | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 668, in step | |
getattr(self, inst.opname)(inst) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 1251, in BUILD_SLICE | |
SliceVariable( | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/base.py", line 27, in __call__ | |
obj = type.__call__(cls, *args, **kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/lists.py", line 616, in __init__ | |
unimplemented("Dynamic slicing on data-dependent value is not supported") | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/exc.py", line 140, in unimplemented | |
raise Unsupported(msg) | |
torch._dynamo.exc.Unsupported: Dynamic slicing on data-dependent value is not supported | |
[2023-07-06 04:20:53,576] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:53,576] torch._dynamo.output_graph: [DEBUG] COMPILING GRAPH due to GraphCompileReason(reason='step_unsupported', user_stack=[<FrameSummary file /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py, line 218 in forward>], graph_break=True) | |
[2023-07-06 04:20:53,578] torch._dynamo.output_graph: [DEBUG] REMOVE UNUSED GRAPHARG L['attention_mask'] | |
[2023-07-06 04:20:53,578] torch._dynamo.output_graph: [DEBUG] REMOVE UNUSED GRAPHARG L['past_key_value'] | |
[2023-07-06 04:20:53,580] torch._dynamo.output_graph.__graph_code: [DEBUG] TRACED GRAPH | |
===== __compiled_fn_7 ===== | |
<eval_with_key>.139 class GraphModule(torch.nn.Module): | |
def forward(self, L_hidden_states_ : torch.Tensor, L_position_ids_ : torch.Tensor, L_valid_past_index_ : torch.Tensor): | |
l_hidden_states_ = L_hidden_states_ | |
l_position_ids_ = L_position_ids_ | |
l_valid_past_index_ = L_valid_past_index_ | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:205, code: query_key_value_states = self.qkv_proj(hidden_states).view(bsz, q_len, 3, self.num_heads, self.head_dim).permute(2, 0, 3, 1, 4) | |
l__self___qkv_proj = self.L__self___qkv_proj(l_hidden_states_); l_hidden_states_ = None | |
view = l__self___qkv_proj.view(1, 1000, 3, 4, 4); l__self___qkv_proj = None | |
permute = view.permute(2, 0, 3, 1, 4); view = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:207, code: query_states = query_key_value_states[0] | |
getitem = permute[0] | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:208, code: key_value_states = query_key_value_states[1:] | |
getitem_1 = permute[slice(1, None, None)]; permute = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:211, code: cos, sin = self.rotary_emb(key_value_states, seq_len=valid_past_index + q_len) | |
add = l_valid_past_index_ + 1000 | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:122, code: self.cos_cached.to(dtype=x.dtype), | |
l__self___rotary_emb_cos_cached = self.L__self___rotary_emb_cos_cached | |
to = l__self___rotary_emb_cos_cached.to(dtype = torch.float16); l__self___rotary_emb_cos_cached = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:123, code: self.sin_cached.to(dtype=x.dtype), | |
l__self___rotary_emb_sin_cached = self.L__self___rotary_emb_sin_cached | |
to_1 = l__self___rotary_emb_sin_cached.to(dtype = torch.float16); l__self___rotary_emb_sin_cached = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:212, code: query_states = apply_rotary_pos_emb_opt(query_states, key_value_states[0], cos, sin, position_ids) | |
getitem_2 = getitem_1[0] | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:139, code: cos = cos.squeeze(1).squeeze(0) # [seq_len, dim] | |
squeeze = to.squeeze(1); to = None | |
squeeze_1 = squeeze.squeeze(0); squeeze = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:140, code: sin = sin.squeeze(1).squeeze(0) # [seq_len, dim] | |
squeeze_2 = to_1.squeeze(1); to_1 = None | |
squeeze_3 = squeeze_2.squeeze(0); squeeze_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:141, code: cos = cos[position_ids].unsqueeze(1) # [bs, 1, seq_len, dim] | |
getitem_3 = squeeze_1[l_position_ids_]; squeeze_1 = None | |
unsqueeze = getitem_3.unsqueeze(1); getitem_3 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:142, code: sin = sin[position_ids].unsqueeze(1) # [bs, 1, seq_len, dim] | |
getitem_4 = squeeze_3[l_position_ids_]; squeeze_3 = l_position_ids_ = None | |
unsqueeze_1 = getitem_4.unsqueeze(1); getitem_4 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:143, code: q_embed = (q * cos) + (rotate_half(q) * sin) | |
mul = getitem * unsqueeze | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:129, code: x = torch.roll(x, shifts=x.shape[-1] // 2, dims=-1) | |
roll = torch.roll(getitem, shifts = 2, dims = -1); getitem = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:130, code: x[..., : x.shape[-1] // 2] *= -1 | |
getitem_5 = roll[(Ellipsis, slice(None, 2, None))] | |
getitem_5 *= -1; imul = getitem_5; getitem_5 = None | |
roll[(Ellipsis, slice(None, 2, None))] = imul; setitem = roll; imul = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:143, code: q_embed = (q * cos) + (rotate_half(q) * sin) | |
mul_1 = roll * unsqueeze_1; roll = None | |
add_1 = mul + mul_1; mul = mul_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:144, code: key_states.copy_((key_states * cos) + (rotate_half(key_states) * sin)) | |
mul_2 = getitem_2 * unsqueeze; unsqueeze = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:129, code: x = torch.roll(x, shifts=x.shape[-1] // 2, dims=-1) | |
roll_1 = torch.roll(getitem_2, shifts = 2, dims = -1) | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:130, code: x[..., : x.shape[-1] // 2] *= -1 | |
getitem_6 = roll_1[(Ellipsis, slice(None, 2, None))] | |
getitem_6 *= -1; imul_1 = getitem_6; getitem_6 = None | |
roll_1[(Ellipsis, slice(None, 2, None))] = imul_1; setitem_1 = roll_1; imul_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:144, code: key_states.copy_((key_states * cos) + (rotate_half(key_states) * sin)) | |
mul_3 = roll_1 * unsqueeze_1; roll_1 = unsqueeze_1 = None | |
add_2 = mul_2 + mul_3; mul_2 = mul_3 = None | |
copy_ = getitem_2.copy_(add_2); getitem_2 = add_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:216, code: past_kv_slice_end = torch.eq(valid_past_index, 0).int() * q_len + torch.ne(valid_past_index, 0).int() * (valid_past_index + 1) | |
eq = torch.eq(l_valid_past_index_, 0) | |
int_1 = eq.int(); eq = None | |
mul_4 = int_1 * 1000; int_1 = None | |
ne = torch.ne(l_valid_past_index_, 0) | |
int_2 = ne.int(); ne = None | |
add_3 = l_valid_past_index_ + 1 | |
mul_5 = int_2 * add_3; int_2 = add_3 = None | |
add_4 = mul_4 + mul_5; mul_4 = mul_5 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:217, code: past_state_slice_end = torch.eq(valid_past_index, 0).int() * key_value_states.shape[-2] + torch.ne(valid_past_index, 0).int() * (past_kv_slice_end) | |
eq_1 = torch.eq(l_valid_past_index_, 0) | |
int_3 = eq_1.int(); eq_1 = None | |
mul_6 = int_3 * 1000; int_3 = None | |
ne_1 = torch.ne(l_valid_past_index_, 0); l_valid_past_index_ = None | |
int_4 = ne_1.int(); ne_1 = None | |
mul_7 = int_4 * add_4; int_4 = None | |
add_5 = mul_6 + mul_7; mul_6 = mul_7 = None | |
return (add_1, getitem_1, add_4, add_5) | |
[2023-07-06 04:20:53,581] torch._dynamo.output_graph.__graph: [DEBUG] TRACED GRAPH | |
__compiled_fn_7 <eval_with_key>.139 opcode name target args kwargs | |
------------- ------------------------------- ------------------------------------------------------- -------------------------------------------------- ------------------------- | |
placeholder l_hidden_states_ L_hidden_states_ () {} | |
placeholder l_position_ids_ L_position_ids_ () {} | |
placeholder l_valid_past_index_ L_valid_past_index_ () {} | |
call_module l__self___qkv_proj L__self___qkv_proj (l_hidden_states_,) {} | |
call_method view view (l__self___qkv_proj, 1, 1000, 3, 4, 4) {} | |
call_method permute permute (view, 2, 0, 3, 1, 4) {} | |
call_function getitem <built-in function getitem> (permute, 0) {} | |
call_function getitem_1 <built-in function getitem> (permute, slice(1, None, None)) {} | |
call_function add <built-in function add> (l_valid_past_index_, 1000) {} | |
get_attr l__self___rotary_emb_cos_cached L__self___rotary_emb_cos_cached () {} | |
call_method to to (l__self___rotary_emb_cos_cached,) {'dtype': torch.float16} | |
get_attr l__self___rotary_emb_sin_cached L__self___rotary_emb_sin_cached () {} | |
call_method to_1 to (l__self___rotary_emb_sin_cached,) {'dtype': torch.float16} | |
call_function getitem_2 <built-in function getitem> (getitem_1, 0) {} | |
call_method squeeze squeeze (to, 1) {} | |
call_method squeeze_1 squeeze (squeeze, 0) {} | |
call_method squeeze_2 squeeze (to_1, 1) {} | |
call_method squeeze_3 squeeze (squeeze_2, 0) {} | |
call_function getitem_3 <built-in function getitem> (squeeze_1, l_position_ids_) {} | |
call_method unsqueeze unsqueeze (getitem_3, 1) {} | |
call_function getitem_4 <built-in function getitem> (squeeze_3, l_position_ids_) {} | |
call_method unsqueeze_1 unsqueeze (getitem_4, 1) {} | |
call_function mul <built-in function mul> (getitem, unsqueeze) {} | |
call_function roll <built-in method roll of type object at 0x7fb940469580> (getitem,) {'shifts': 2, 'dims': -1} | |
call_function getitem_5 <built-in function getitem> (roll, (Ellipsis, slice(None, 2, None))) {} | |
call_function imul <built-in function imul> (getitem_5, -1) {} | |
call_function setitem <built-in function setitem> (roll, (Ellipsis, slice(None, 2, None)), imul) {} | |
call_function mul_1 <built-in function mul> (roll, unsqueeze_1) {} | |
call_function add_1 <built-in function add> (mul, mul_1) {} | |
call_function mul_2 <built-in function mul> (getitem_2, unsqueeze) {} | |
call_function roll_1 <built-in method roll of type object at 0x7fb940469580> (getitem_2,) {'shifts': 2, 'dims': -1} | |
call_function getitem_6 <built-in function getitem> (roll_1, (Ellipsis, slice(None, 2, None))) {} | |
call_function imul_1 <built-in function imul> (getitem_6, -1) {} | |
call_function setitem_1 <built-in function setitem> (roll_1, (Ellipsis, slice(None, 2, None)), imul_1) {} | |
call_function mul_3 <built-in function mul> (roll_1, unsqueeze_1) {} | |
call_function add_2 <built-in function add> (mul_2, mul_3) {} | |
call_method copy_ copy_ (getitem_2, add_2) {} | |
call_function eq <built-in method eq of type object at 0x7fb940469580> (l_valid_past_index_, 0) {} | |
call_method int_1 int (eq,) {} | |
call_function mul_4 <built-in function mul> (int_1, 1000) {} | |
call_function ne <built-in method ne of type object at 0x7fb940469580> (l_valid_past_index_, 0) {} | |
call_method int_2 int (ne,) {} | |
call_function add_3 <built-in function add> (l_valid_past_index_, 1) {} | |
call_function mul_5 <built-in function mul> (int_2, add_3) {} | |
call_function add_4 <built-in function add> (mul_4, mul_5) {} | |
call_function eq_1 <built-in method eq of type object at 0x7fb940469580> (l_valid_past_index_, 0) {} | |
call_method int_3 int (eq_1,) {} | |
call_function mul_6 <built-in function mul> (int_3, 1000) {} | |
call_function ne_1 <built-in method ne of type object at 0x7fb940469580> (l_valid_past_index_, 0) {} | |
call_method int_4 int (ne_1,) {} | |
call_function mul_7 <built-in function mul> (int_4, add_4) {} | |
call_function add_5 <built-in function add> (mul_6, mul_7) {} | |
output output output ((add_1, getitem_1, add_4, add_5),) {} | |
[2023-07-06 04:20:53,583] torch._dynamo.output_graph: [INFO] Step 2: calling compiler function inductor | |
[2023-07-06 04:20:53,705] torch._functorch.aot_autograd.__aot_graphs: [INFO] TRACED GRAPH | |
===== Forward graph 17 ===== | |
<eval_with_key>.143 class <lambda>(torch.nn.Module): | |
def forward(self, arg0_1: f16[48, 16], arg1_1: f32[1, 1, 2048, 4], arg2_1: f32[1, 1, 2048, 4], arg3_1: f16[1, 1000, 16], arg4_1: i64[1, 1000], arg5_1: i64[]): | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:205, code: query_key_value_states = self.qkv_proj(hidden_states).view(bsz, q_len, 3, self.num_heads, self.head_dim).permute(2, 0, 3, 1, 4) | |
permute: f16[16, 48] = torch.ops.aten.permute.default(arg0_1, [1, 0]); arg0_1 = None | |
view: f16[1000, 16] = torch.ops.aten.view.default(arg3_1, [1000, 16]); arg3_1 = None | |
mm: f16[1000, 48] = torch.ops.aten.mm.default(view, permute); view = permute = None | |
view_1: f16[1, 1000, 48] = torch.ops.aten.view.default(mm, [1, 1000, 48]) | |
view_2: f16[1, 1000, 3, 4, 4] = torch.ops.aten.view.default(view_1, [1, 1000, 3, 4, 4]); view_1 = None | |
permute_1: f16[3, 1, 4, 1000, 4] = torch.ops.aten.permute.default(view_2, [2, 0, 3, 1, 4]); view_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:207, code: query_states = query_key_value_states[0] | |
select: f16[1, 4, 1000, 4] = torch.ops.aten.select.int(permute_1, 0, 0) | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:208, code: key_value_states = query_key_value_states[1:] | |
slice_1: f16[2, 1, 4, 1000, 4] = torch.ops.aten.slice.Tensor(permute_1, 0, 1, 9223372036854775807); permute_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:122, code: self.cos_cached.to(dtype=x.dtype), | |
convert_element_type: f16[1, 1, 2048, 4] = torch.ops.prims.convert_element_type.default(arg1_1, torch.float16); arg1_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:123, code: self.sin_cached.to(dtype=x.dtype), | |
convert_element_type_1: f16[1, 1, 2048, 4] = torch.ops.prims.convert_element_type.default(arg2_1, torch.float16); arg2_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:212, code: query_states = apply_rotary_pos_emb_opt(query_states, key_value_states[0], cos, sin, position_ids) | |
select_1: f16[1, 4, 1000, 4] = torch.ops.aten.select.int(slice_1, 0, 0); slice_1 = None | |
# No stacktrace found for following nodes | |
squeeze: f16[1, 2048, 4] = torch.ops.aten.squeeze.dim(convert_element_type, 1); convert_element_type = None | |
squeeze_1: f16[2048, 4] = torch.ops.aten.squeeze.dim(squeeze, 0); squeeze = None | |
squeeze_2: f16[1, 2048, 4] = torch.ops.aten.squeeze.dim(convert_element_type_1, 1); convert_element_type_1 = None | |
squeeze_3: f16[2048, 4] = torch.ops.aten.squeeze.dim(squeeze_2, 0); squeeze_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:141, code: cos = cos[position_ids].unsqueeze(1) # [bs, 1, seq_len, dim] | |
index: f16[1, 1000, 4] = torch.ops.aten.index.Tensor(squeeze_1, [arg4_1]); squeeze_1 = None | |
unsqueeze: f16[1, 1, 1000, 4] = torch.ops.aten.unsqueeze.default(index, 1); index = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:142, code: sin = sin[position_ids].unsqueeze(1) # [bs, 1, seq_len, dim] | |
index_1: f16[1, 1000, 4] = torch.ops.aten.index.Tensor(squeeze_3, [arg4_1]); squeeze_3 = arg4_1 = None | |
unsqueeze_1: f16[1, 1, 1000, 4] = torch.ops.aten.unsqueeze.default(index_1, 1); index_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:143, code: q_embed = (q * cos) + (rotate_half(q) * sin) | |
mul: f16[1, 4, 1000, 4] = torch.ops.aten.mul.Tensor(select, unsqueeze) | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:129, code: x = torch.roll(x, shifts=x.shape[-1] // 2, dims=-1) | |
roll: f16[1, 4, 1000, 4] = torch.ops.aten.roll.default(select, [2], [-1]); select = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:130, code: x[..., : x.shape[-1] // 2] *= -1 | |
slice_2: f16[1, 4, 1000, 2] = torch.ops.aten.slice.Tensor(roll, 3, 0, 2) | |
mul_1: f16[1, 4, 1000, 2] = torch.ops.aten.mul.Tensor(slice_2, -1); slice_2 = None | |
slice_scatter: f16[1, 4, 1000, 4] = torch.ops.aten.slice_scatter.default(roll, mul_1, 3, 0, 2); roll = mul_1 = None | |
slice_3: f16[1, 4, 1000, 2] = torch.ops.aten.slice.Tensor(slice_scatter, 3, 0, 2) | |
slice_scatter_1: f16[1, 4, 1000, 4] = torch.ops.aten.slice_scatter.default(slice_scatter, slice_3, 3, 0, 2); slice_scatter = slice_3 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:143, code: q_embed = (q * cos) + (rotate_half(q) * sin) | |
mul_2: f16[1, 4, 1000, 4] = torch.ops.aten.mul.Tensor(slice_scatter_1, unsqueeze_1); slice_scatter_1 = None | |
add_1: f16[1, 4, 1000, 4] = torch.ops.aten.add.Tensor(mul, mul_2); mul = mul_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:144, code: key_states.copy_((key_states * cos) + (rotate_half(key_states) * sin)) | |
mul_3: f16[1, 4, 1000, 4] = torch.ops.aten.mul.Tensor(select_1, unsqueeze); unsqueeze = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:129, code: x = torch.roll(x, shifts=x.shape[-1] // 2, dims=-1) | |
roll_1: f16[1, 4, 1000, 4] = torch.ops.aten.roll.default(select_1, [2], [-1]); select_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:130, code: x[..., : x.shape[-1] // 2] *= -1 | |
slice_7: f16[1, 4, 1000, 2] = torch.ops.aten.slice.Tensor(roll_1, 3, 0, 2) | |
mul_4: f16[1, 4, 1000, 2] = torch.ops.aten.mul.Tensor(slice_7, -1); slice_7 = None | |
slice_scatter_2: f16[1, 4, 1000, 4] = torch.ops.aten.slice_scatter.default(roll_1, mul_4, 3, 0, 2); roll_1 = mul_4 = None | |
slice_8: f16[1, 4, 1000, 2] = torch.ops.aten.slice.Tensor(slice_scatter_2, 3, 0, 2) | |
slice_scatter_3: f16[1, 4, 1000, 4] = torch.ops.aten.slice_scatter.default(slice_scatter_2, slice_8, 3, 0, 2); slice_scatter_2 = slice_8 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:144, code: key_states.copy_((key_states * cos) + (rotate_half(key_states) * sin)) | |
mul_5: f16[1, 4, 1000, 4] = torch.ops.aten.mul.Tensor(slice_scatter_3, unsqueeze_1); slice_scatter_3 = unsqueeze_1 = None | |
add_2: f16[1, 4, 1000, 4] = torch.ops.aten.add.Tensor(mul_3, mul_5); mul_3 = mul_5 = None | |
view_3: f16[1, 1000, 48] = torch.ops.aten.view.default(mm, [1, 1000, 48]); mm = None | |
view_4: f16[1, 1000, 3, 4, 4] = torch.ops.aten.view.default(view_3, [1, 1000, 3, 4, 4]); view_3 = None | |
permute_2: f16[3, 1, 4, 1000, 4] = torch.ops.aten.permute.default(view_4, [2, 0, 3, 1, 4]); view_4 = None | |
slice_12: f16[2, 1, 4, 1000, 4] = torch.ops.aten.slice.Tensor(permute_2, 0, 1, 9223372036854775807) | |
select_scatter: f16[2, 1, 4, 1000, 4] = torch.ops.aten.select_scatter.default(slice_12, add_2, 0, 0); slice_12 = add_2 = None | |
slice_scatter_4: f16[3, 1, 4, 1000, 4] = torch.ops.aten.slice_scatter.default(permute_2, select_scatter, 0, 1, 9223372036854775807); permute_2 = select_scatter = None | |
permute_3: f16[1, 1000, 3, 4, 4] = torch.ops.aten.permute.default(slice_scatter_4, [1, 3, 0, 2, 4]); slice_scatter_4 = None | |
view_5: f16[1, 1000, 48] = torch.ops.aten.view.default(permute_3, [1, 1000, 48]); permute_3 = None | |
view_6: f16[1000, 48] = torch.ops.aten.view.default(view_5, [1000, 48]); view_5 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:216, code: past_kv_slice_end = torch.eq(valid_past_index, 0).int() * q_len + torch.ne(valid_past_index, 0).int() * (valid_past_index + 1) | |
eq: b8[] = torch.ops.aten.eq.Scalar(arg5_1, 0) | |
convert_element_type_2: i32[] = torch.ops.prims.convert_element_type.default(eq, torch.int32); eq = None | |
mul_6: i32[] = torch.ops.aten.mul.Tensor(convert_element_type_2, 1000); convert_element_type_2 = None | |
ne: b8[] = torch.ops.aten.ne.Scalar(arg5_1, 0) | |
convert_element_type_3: i32[] = torch.ops.prims.convert_element_type.default(ne, torch.int32); ne = None | |
add_3: i64[] = torch.ops.aten.add.Tensor(arg5_1, 1) | |
mul_7: i64[] = torch.ops.aten.mul.Tensor(convert_element_type_3, add_3); convert_element_type_3 = add_3 = None | |
add_4: i64[] = torch.ops.aten.add.Tensor(mul_6, mul_7); mul_6 = mul_7 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:217, code: past_state_slice_end = torch.eq(valid_past_index, 0).int() * key_value_states.shape[-2] + torch.ne(valid_past_index, 0).int() * (past_kv_slice_end) | |
eq_1: b8[] = torch.ops.aten.eq.Scalar(arg5_1, 0) | |
convert_element_type_4: i32[] = torch.ops.prims.convert_element_type.default(eq_1, torch.int32); eq_1 = None | |
mul_8: i32[] = torch.ops.aten.mul.Tensor(convert_element_type_4, 1000); convert_element_type_4 = None | |
ne_1: b8[] = torch.ops.aten.ne.Scalar(arg5_1, 0); arg5_1 = None | |
convert_element_type_5: i32[] = torch.ops.prims.convert_element_type.default(ne_1, torch.int32); ne_1 = None | |
mul_9: i64[] = torch.ops.aten.mul.Tensor(convert_element_type_5, add_4); convert_element_type_5 = None | |
add_5: i64[] = torch.ops.aten.add.Tensor(mul_8, mul_9); mul_8 = mul_9 = None | |
# No stacktrace found for following nodes | |
view_9: f16[1, 1000, 48] = torch.ops.aten.view.default(view_6, [1, 1000, 48]); view_6 = None | |
view_10: f16[1, 1000, 3, 4, 4] = torch.ops.aten.view.default(view_9, [1, 1000, 3, 4, 4]); view_9 = None | |
permute_5: f16[3, 1, 4, 1000, 4] = torch.ops.aten.permute.default(view_10, [2, 0, 3, 1, 4]); view_10 = None | |
slice_14: f16[2, 1, 4, 1000, 4] = torch.ops.aten.slice.Tensor(permute_5, 0, 1, 9223372036854775807); permute_5 = None | |
return (add_1, slice_14, add_4, add_5) | |
[2023-07-06 04:20:53,734] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 4 | |
[2023-07-06 04:20:53,860] torch._inductor.ir: [DEBUG] convert_to_reinterpret_view failed: stride=[16000, 0, 4000, 4, 1] offset=16000 index=16000*r0 + 4000*r1 + 4*r2 + r3 + 16000*(((4*r1 + r3)//16)) + 16000 | |
[2023-07-06 04:20:53,866] torch._inductor.graph: [DEBUG] Force channels last inputs for 0 conv for the current graph with id 4 | |
[2023-07-06 04:20:53,989] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_ubuntu/lo/clozylkw544bb7yqlyamdizaia7jo2uesdfqh4qhrfm2er5ahkkn.py | |
[2023-07-06 04:20:53,989] torch._inductor.compile_fx: [DEBUG] disabled cudagraphs because ['non-cuda device in graph'] | |
skipping cudagraphs due to multiple devices | |
[2023-07-06 04:20:53,989] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 4 | |
[2023-07-06 04:20:53,991] torch._dynamo.output_graph: [INFO] Step 2: done compiler function inductor | |
[2023-07-06 04:20:54,000] torch._dynamo.convert_frame.__guards: [DEBUG] GUARDS: | |
___check_obj_id(L['self'], 140431662491744) | |
hasattr(L['position_ids'], '_dynamo_dynamic_indices') == False | |
hasattr(L['hidden_states'], '_dynamo_dynamic_indices') == False | |
hasattr(L['attention_mask'], '_dynamo_dynamic_indices') == False | |
hasattr(L['past_key_value'], '_dynamo_dynamic_indices') == False | |
hasattr(L['valid_past_index'], '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['output_attentions'], 7698112) | |
not ___is_grad_enabled() | |
not ___are_deterministic_algorithms_enabled() | |
utils_device.CURRENT_DEVICE == None | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks.keys()) == set() | |
[2023-07-06 04:20:54,001] torch._dynamo.convert_frame.__recompiles: [DEBUG] ('Recompiling function <resume in forward> in /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py', 'set env var TORCHDYNAMO_REPORT_GUARD_FAILURES=1 to debug further') | |
[2023-07-06 04:20:54,002] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo start tracing <resume in forward> /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:286 | |
[2023-07-06 04:20:54,002] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['___stack0'][0] (1, 1000, 16) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None] | |
[2023-07-06 04:20:54,004] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['residual'] (1, 1000, 16) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None] | |
[2023-07-06 04:20:54,005] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST ___stack0 [] | |
[2023-07-06 04:20:54,005] torch._dynamo.symbolic_convert: [DEBUG] TRACE JUMP_ABSOLUTE 38 [TupleVariable()] | |
[2023-07-06 04:20:54,005] torch._dynamo.symbolic_convert: [DEBUG] TRACE UNPACK_SEQUENCE 3 [TupleVariable()] | |
[2023-07-06 04:20:54,005] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [ConstantVariable(NoneType), ConstantVariable(NoneType), TensorVariable()] | |
[2023-07-06 04:20:54,005] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST self_attn_weights [ConstantVariable(NoneType), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:54,005] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST present_key_value [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:54,005] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:294 | |
[2023-07-06 04:20:54,005] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST residual [] | |
[2023-07-06 04:20:54,005] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:54,005] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:54,006] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:54,006] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:297 | |
[2023-07-06 04:20:54,006] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:54,006] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST residual [TensorVariable()] | |
[2023-07-06 04:20:54,006] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:298 | |
[2023-07-06 04:20:54,006] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,006] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR post_attention_layernorm [NNModuleVariable()] | |
[2023-07-06 04:20:54,007] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [NNModuleVariable()] | |
[2023-07-06 04:20:54,007] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:54,007] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:54,008] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:54,009] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:54,009] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:54,009] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:54,009] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:54,009] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:54,009] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,009] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:54,010] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable())] | |
[2023-07-06 04:20:54,010] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,010] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,010] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:54,010] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,010] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,010] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:54,011] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,011] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,011] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:54,011] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,011] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,011] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:54,011] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,011] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:54,011] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:54,012] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,012] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,012] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:54,012] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:54,012] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,012] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,012] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:54,012] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:54,012] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,012] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,012] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:54,012] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:54,012] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,012] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,012] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:54,012] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:54,012] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable())] | |
[2023-07-06 04:20:54,012] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:54,013] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:54,013] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:54,013] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:54,013] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object forward at 0x7fb8cddd7ec0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 83> | |
[2023-07-06 04:20:54,013] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:84 | |
[2023-07-06 04:20:54,013] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:54,013] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR dtype [TensorVariable()] | |
[2023-07-06 04:20:54,013] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST input_dtype [TorchVariable(torch.float16)] | |
[2023-07-06 04:20:54,013] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85 | |
[2023-07-06 04:20:54,014] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:54,014] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable()] | |
[2023-07-06 04:20:54,014] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:54,014] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR float32 [GetAttrVariable(TensorVariable(), to), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:54,014] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), to), ConstantVariable(dtype)] | |
[2023-07-06 04:20:54,015] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR pow [TensorVariable()] | |
[2023-07-06 04:20:54,016] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [GetAttrVariable(TensorVariable(), pow)] | |
[2023-07-06 04:20:54,016] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), pow), ConstantVariable(int)] | |
[2023-07-06 04:20:54,017] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR mean [TensorVariable()] | |
[2023-07-06 04:20:54,017] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [GetAttrVariable(TensorVariable(), mean)] | |
[2023-07-06 04:20:54,017] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST True [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int)] | |
[2023-07-06 04:20:54,017] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('keepdim',) [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int), ConstantVariable(bool)] | |
[2023-07-06 04:20:54,017] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 2 [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int), ConstantVariable(bool), ConstantVariable(tuple)] | |
[2023-07-06 04:20:54,018] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST variance [TensorVariable()] | |
[2023-07-06 04:20:54,018] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86 | |
[2023-07-06 04:20:54,018] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:54,018] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [TensorVariable()] | |
[2023-07-06 04:20:54,018] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR rsqrt [TensorVariable(), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:54,018] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST variance [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:54,019] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:54,019] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR variance_epsilon [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:54,019] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(float)] | |
[2023-07-06 04:20:54,020] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:54,020] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:54,021] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:54,021] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88 | |
[2023-07-06 04:20:54,021] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,021] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR weight [NNModuleVariable()] | |
[2023-07-06 04:20:54,022] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['self'].post_attention_layernorm.weight (16,) [<DimDynamic.STATIC: 2>] [None] | |
[2023-07-06 04:20:54,022] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:54,022] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:54,023] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable()] | |
[2023-07-06 04:20:54,023] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST input_dtype [GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:54,023] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16)] | |
[2023-07-06 04:20:54,024] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:54,025] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object forward at 0x7fb8cddd7ec0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 83> | |
[2023-07-06 04:20:54,025] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:54,025] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:54,025] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:54,025] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:299 | |
[2023-07-06 04:20:54,025] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,025] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR mlp [NNModuleVariable()] | |
[2023-07-06 04:20:54,025] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [NNModuleVariable()] | |
[2023-07-06 04:20:54,025] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:54,026] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:54,027] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:54,027] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:54,027] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:54,027] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:54,028] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:54,028] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:54,028] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,028] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:54,028] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function LlamaMLP.forward at 0x7fb8d0c464d0>, NNModuleVariable())] | |
[2023-07-06 04:20:54,028] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,028] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,028] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:54,028] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,029] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,029] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:54,029] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,029] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,029] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:54,029] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,029] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,029] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:54,030] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,030] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:54,030] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:54,030] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,030] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,030] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:54,030] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:54,030] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,030] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,030] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:54,030] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:54,030] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,030] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,030] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:54,030] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:54,030] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,030] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,030] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:54,030] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:54,030] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function LlamaMLP.forward at 0x7fb8d0c464d0>, NNModuleVariable())] | |
[2023-07-06 04:20:54,030] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function LlamaMLP.forward at 0x7fb8d0c464d0>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:54,030] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function LlamaMLP.forward at 0x7fb8d0c464d0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:54,030] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function LlamaMLP.forward at 0x7fb8d0c464d0>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:54,031] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function LlamaMLP.forward at 0x7fb8d0c464d0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:54,031] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object forward at 0x7fb8cddf85b0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 161> | |
[2023-07-06 04:20:54,031] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:162 | |
[2023-07-06 04:20:54,031] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,031] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR down_proj [NNModuleVariable()] | |
[2023-07-06 04:20:54,031] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [NNModuleVariable()] | |
[2023-07-06 04:20:54,031] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR act_fn [NNModuleVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:54,032] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [NNModuleVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:54,032] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR gate_proj [NNModuleVariable(), NNModuleVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:54,032] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [NNModuleVariable(), NNModuleVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:54,032] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), NNModuleVariable(), NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:54,036] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:54,036] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:54,037] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:54,037] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:54,037] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:54,038] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:54,038] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:54,038] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:54,038] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,038] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:54,038] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function SiLUActivation.forward at 0x7fb8ccf64280>, NNModuleVariable())] | |
[2023-07-06 04:20:54,038] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,039] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,039] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:54,039] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,039] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,039] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:54,039] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,039] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,039] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:54,040] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,040] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,040] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:54,040] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,040] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:54,040] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:54,040] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,040] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,040] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:54,040] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:54,040] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,040] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,040] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:54,040] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:54,040] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,040] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,040] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:54,040] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:54,040] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,040] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,041] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:54,041] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:54,041] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function SiLUActivation.forward at 0x7fb8ccf64280>, NNModuleVariable())] | |
[2023-07-06 04:20:54,041] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function SiLUActivation.forward at 0x7fb8ccf64280>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:54,041] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function SiLUActivation.forward at 0x7fb8ccf64280>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:54,041] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function SiLUActivation.forward at 0x7fb8ccf64280>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:54,041] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function SiLUActivation.forward at 0x7fb8ccf64280>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:54,041] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object forward at 0x7fb8ccf62760, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/activations.py", line 149> | |
[2023-07-06 04:20:54,041] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/activations.py:150 | |
[2023-07-06 04:20:54,041] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL nn [] | |
[2023-07-06 04:20:54,041] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR functional [TorchVariable(<module 'torch.nn' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/__init__.py'>)] | |
[2023-07-06 04:20:54,042] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR silu [TorchVariable(<module 'torch.nn.functional' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/functional.py'>)] | |
[2023-07-06 04:20:54,042] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST input [TorchVariable(<function silu at 0x7fb945bc5990>)] | |
[2023-07-06 04:20:54,042] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [TorchVariable(<function silu at 0x7fb945bc5990>), TensorVariable()] | |
[2023-07-06 04:20:54,045] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:54,045] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object forward at 0x7fb8ccf62760, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/activations.py", line 149> | |
[2023-07-06 04:20:54,045] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:54,045] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:54,045] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:54,045] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR up_proj [NNModuleVariable(), TensorVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:54,046] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [NNModuleVariable(), TensorVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:54,046] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable(), NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:54,048] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [NNModuleVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:54,049] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:54,052] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:54,052] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object forward at 0x7fb8cddf85b0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 161> | |
[2023-07-06 04:20:54,052] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:54,052] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:54,052] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:54,052] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:300 | |
[2023-07-06 04:20:54,052] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST residual [] | |
[2023-07-06 04:20:54,052] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:54,052] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:54,053] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:54,053] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:302 | |
[2023-07-06 04:20:54,053] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:54,053] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_TUPLE 1 [TensorVariable()] | |
[2023-07-06 04:20:54,053] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST outputs [TupleVariable()] | |
[2023-07-06 04:20:54,053] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:304 | |
[2023-07-06 04:20:54,053] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST output_attentions [] | |
[2023-07-06 04:20:54,053] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 106 [ConstantVariable(bool)] | |
[2023-07-06 04:20:54,053] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:307 | |
[2023-07-06 04:20:54,053] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST outputs [] | |
[2023-07-06 04:20:54,053] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST present_key_value [TupleVariable()] | |
[2023-07-06 04:20:54,053] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_TUPLE 1 [TupleVariable(), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:54,054] torch._dynamo.symbolic_convert: [DEBUG] TRACE INPLACE_ADD None [TupleVariable(), TupleVariable()] | |
[2023-07-06 04:20:54,054] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST outputs [TupleVariable()] | |
[2023-07-06 04:20:54,054] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:309 | |
[2023-07-06 04:20:54,054] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST outputs [] | |
[2023-07-06 04:20:54,054] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TupleVariable()] | |
[2023-07-06 04:20:54,054] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo done tracing <resume in forward> (RETURN_VALUE) | |
[2023-07-06 04:20:54,054] torch._dynamo.symbolic_convert: [DEBUG] RETURN_VALUE triggered compile | |
[2023-07-06 04:20:54,054] torch._dynamo.output_graph: [DEBUG] COMPILING GRAPH due to GraphCompileReason(reason='return_value', user_stack=[<FrameSummary file /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py, line 309 in <resume in forward>>], graph_break=False) | |
[2023-07-06 04:20:54,056] torch._dynamo.output_graph.__graph_code: [DEBUG] TRACED GRAPH | |
===== __compiled_fn_8 ===== | |
<eval_with_key>.148 class GraphModule(torch.nn.Module): | |
def forward(self, L_stack0_0_ : torch.Tensor, L_residual_ : torch.Tensor): | |
l_stack0_0_ = L_stack0_0_ | |
l_residual_ = L_residual_ | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:294, code: hidden_states = residual + hidden_states | |
add = l_residual_ + l_stack0_0_; l_residual_ = l_stack0_0_ = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85, code: variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) | |
to = add.to(torch.float32) | |
pow_1 = to.pow(2); to = None | |
mean = pow_1.mean(-1, keepdim = True); pow_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86, code: hidden_states = hidden_states * torch.rsqrt(variance + self.variance_epsilon) | |
add_1 = mean + 1e-06; mean = None | |
rsqrt = torch.rsqrt(add_1); add_1 = None | |
mul = add * rsqrt; rsqrt = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88, code: return (self.weight * hidden_states).to(input_dtype) | |
l__self___post_attention_layernorm_weight = self.L__self___post_attention_layernorm_weight | |
mul_1 = l__self___post_attention_layernorm_weight * mul; l__self___post_attention_layernorm_weight = mul = None | |
to_1 = mul_1.to(torch.float16); mul_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:162, code: return self.down_proj(self.act_fn(self.gate_proj(x)) * self.up_proj(x)) | |
l__self___mlp_gate_proj = self.L__self___mlp_gate_proj(to_1) | |
# File: /home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/activations.py:150, code: return nn.functional.silu(input) | |
silu = torch.nn.functional.silu(l__self___mlp_gate_proj); l__self___mlp_gate_proj = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:162, code: return self.down_proj(self.act_fn(self.gate_proj(x)) * self.up_proj(x)) | |
l__self___mlp_up_proj = self.L__self___mlp_up_proj(to_1); to_1 = None | |
mul_2 = silu * l__self___mlp_up_proj; silu = l__self___mlp_up_proj = None | |
l__self___mlp_down_proj = self.L__self___mlp_down_proj(mul_2); mul_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:300, code: hidden_states = residual + hidden_states | |
add_2 = add + l__self___mlp_down_proj; add = l__self___mlp_down_proj = None | |
return (add_2,) | |
[2023-07-06 04:20:54,056] torch._dynamo.output_graph.__graph: [DEBUG] TRACED GRAPH | |
__compiled_fn_8 <eval_with_key>.148 opcode name target args kwargs | |
------------- ----------------------------------------- -------------------------------------------------------- ------------------------------------------------ ----------------- | |
placeholder l_stack0_0_ L_stack0_0_ () {} | |
placeholder l_residual_ L_residual_ () {} | |
call_function add <built-in function add> (l_residual_, l_stack0_0_) {} | |
call_method to to (add, torch.float32) {} | |
call_method pow_1 pow (to, 2) {} | |
call_method mean mean (pow_1, -1) {'keepdim': True} | |
call_function add_1 <built-in function add> (mean, 1e-06) {} | |
call_function rsqrt <built-in method rsqrt of type object at 0x7fb940469580> (add_1,) {} | |
call_function mul <built-in function mul> (add, rsqrt) {} | |
get_attr l__self___post_attention_layernorm_weight L__self___post_attention_layernorm_weight () {} | |
call_function mul_1 <built-in function mul> (l__self___post_attention_layernorm_weight, mul) {} | |
call_method to_1 to (mul_1, torch.float16) {} | |
call_module l__self___mlp_gate_proj L__self___mlp_gate_proj (to_1,) {} | |
call_function silu <function silu at 0x7fb945bc5990> (l__self___mlp_gate_proj,) {} | |
call_module l__self___mlp_up_proj L__self___mlp_up_proj (to_1,) {} | |
call_function mul_2 <built-in function mul> (silu, l__self___mlp_up_proj) {} | |
call_module l__self___mlp_down_proj L__self___mlp_down_proj (mul_2,) {} | |
call_function add_2 <built-in function add> (add, l__self___mlp_down_proj) {} | |
output output output ((add_2,),) {} | |
[2023-07-06 04:20:54,057] torch._dynamo.output_graph: [INFO] Step 2: calling compiler function inductor | |
[2023-07-06 04:20:54,105] torch._functorch.aot_autograd.__aot_graphs: [INFO] TRACED GRAPH | |
===== Forward graph 18 ===== | |
<eval_with_key>.152 class <lambda>(torch.nn.Module): | |
def forward(self, arg0_1: f16[16], arg1_1: f16[64, 16], arg2_1: f16[64, 16], arg3_1: f16[16, 64], arg4_1: f16[1, 1000, 16], arg5_1: f16[1, 1000, 16]): | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:294, code: hidden_states = residual + hidden_states | |
add: f16[1, 1000, 16] = torch.ops.aten.add.Tensor(arg5_1, arg4_1); arg5_1 = arg4_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85, code: variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) | |
convert_element_type: f32[1, 1000, 16] = torch.ops.prims.convert_element_type.default(add, torch.float32) | |
pow_1: f32[1, 1000, 16] = torch.ops.aten.pow.Tensor_Scalar(convert_element_type, 2); convert_element_type = None | |
mean: f32[1, 1000, 1] = torch.ops.aten.mean.dim(pow_1, [-1], True); pow_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86, code: hidden_states = hidden_states * torch.rsqrt(variance + self.variance_epsilon) | |
add_1: f32[1, 1000, 1] = torch.ops.aten.add.Tensor(mean, 1e-06); mean = None | |
rsqrt: f32[1, 1000, 1] = torch.ops.aten.rsqrt.default(add_1); add_1 = None | |
mul: f32[1, 1000, 16] = torch.ops.aten.mul.Tensor(add, rsqrt); rsqrt = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88, code: return (self.weight * hidden_states).to(input_dtype) | |
mul_1: f32[1, 1000, 16] = torch.ops.aten.mul.Tensor(arg0_1, mul); arg0_1 = mul = None | |
convert_element_type_1: f16[1, 1000, 16] = torch.ops.prims.convert_element_type.default(mul_1, torch.float16); mul_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:162, code: return self.down_proj(self.act_fn(self.gate_proj(x)) * self.up_proj(x)) | |
permute: f16[16, 64] = torch.ops.aten.permute.default(arg1_1, [1, 0]); arg1_1 = None | |
view: f16[1000, 16] = torch.ops.aten.view.default(convert_element_type_1, [1000, 16]) | |
mm: f16[1000, 64] = torch.ops.aten.mm.default(view, permute); view = permute = None | |
view_1: f16[1, 1000, 64] = torch.ops.aten.view.default(mm, [1, 1000, 64]); mm = None | |
# File: /home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/activations.py:150, code: return nn.functional.silu(input) | |
convert_element_type_2: f32[1, 1000, 64] = torch.ops.prims.convert_element_type.default(view_1, torch.float32); view_1 = None | |
sigmoid: f32[1, 1000, 64] = torch.ops.aten.sigmoid.default(convert_element_type_2) | |
mul_2: f32[1, 1000, 64] = torch.ops.aten.mul.Tensor(convert_element_type_2, sigmoid); convert_element_type_2 = sigmoid = None | |
convert_element_type_3: f16[1, 1000, 64] = torch.ops.prims.convert_element_type.default(mul_2, torch.float16); mul_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:162, code: return self.down_proj(self.act_fn(self.gate_proj(x)) * self.up_proj(x)) | |
permute_1: f16[16, 64] = torch.ops.aten.permute.default(arg2_1, [1, 0]); arg2_1 = None | |
view_2: f16[1000, 16] = torch.ops.aten.view.default(convert_element_type_1, [1000, 16]); convert_element_type_1 = None | |
mm_1: f16[1000, 64] = torch.ops.aten.mm.default(view_2, permute_1); view_2 = permute_1 = None | |
view_3: f16[1, 1000, 64] = torch.ops.aten.view.default(mm_1, [1, 1000, 64]); mm_1 = None | |
mul_3: f16[1, 1000, 64] = torch.ops.aten.mul.Tensor(convert_element_type_3, view_3); convert_element_type_3 = view_3 = None | |
permute_2: f16[64, 16] = torch.ops.aten.permute.default(arg3_1, [1, 0]); arg3_1 = None | |
view_4: f16[1000, 64] = torch.ops.aten.view.default(mul_3, [1000, 64]); mul_3 = None | |
mm_2: f16[1000, 16] = torch.ops.aten.mm.default(view_4, permute_2); view_4 = permute_2 = None | |
view_5: f16[1, 1000, 16] = torch.ops.aten.view.default(mm_2, [1, 1000, 16]); mm_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:300, code: hidden_states = residual + hidden_states | |
add_2: f16[1, 1000, 16] = torch.ops.aten.add.Tensor(add, view_5); add = view_5 = None | |
return (add_2,) | |
[2023-07-06 04:20:54,133] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 5 | |
[2023-07-06 04:20:54,177] torch._inductor.graph: [DEBUG] Force channels last inputs for 0 conv for the current graph with id 5 | |
[2023-07-06 04:20:54,201] torch._inductor.scheduler: [DEBUG] remove_buffer('buf0') | |
[2023-07-06 04:20:54,210] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_ubuntu/or/cor2fodulphywxrtc7zlvp6tsjseyifwlaiv6ticnqodficz2elj.py | |
[2023-07-06 04:20:54,210] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 5 | |
[2023-07-06 04:20:54,212] torch._dynamo.output_graph: [INFO] Step 2: done compiler function inductor | |
[2023-07-06 04:20:54,220] torch._dynamo.convert_frame.__guards: [DEBUG] GUARDS: | |
___check_obj_id(L['self'], 140431662486464) | |
hasattr(L['residual'], '_dynamo_dynamic_indices') == False | |
___check_type_id(L['___stack0'], 7638272) | |
len(L['___stack0']) == 3 | |
hasattr(L['___stack0'][0], '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['___stack0'][1], 7649088) | |
___check_obj_id(L['___stack0'][2], 7649088) | |
___check_obj_id(L['output_attentions'], 7698112) | |
not ___is_grad_enabled() | |
not ___are_deterministic_algorithms_enabled() | |
utils_device.CURRENT_DEVICE == None | |
str(G['torch'].float32) == 'torch.float32' | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks.keys()) == set() | |
[2023-07-06 04:20:54,220] torch._inductor.cudagraph_trees: [DEBUG] Running warmup of function 3 | |
[2023-07-06 04:20:54,221] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo start tracing forward /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:83 | |
[2023-07-06 04:20:54,222] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['hidden_states'] (1, 1000, 16) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None] | |
[2023-07-06 04:20:54,223] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:84 | |
[2023-07-06 04:20:54,223] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:54,223] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR dtype [TensorVariable()] | |
[2023-07-06 04:20:54,226] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST input_dtype [TorchVariable(torch.float16)] | |
[2023-07-06 04:20:54,226] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85 | |
[2023-07-06 04:20:54,226] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:54,226] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable()] | |
[2023-07-06 04:20:54,226] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:54,226] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR float32 [GetAttrVariable(TensorVariable(), to), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:54,226] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), to), ConstantVariable(dtype)] | |
[2023-07-06 04:20:54,228] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR pow [TensorVariable()] | |
[2023-07-06 04:20:54,228] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [GetAttrVariable(TensorVariable(), pow)] | |
[2023-07-06 04:20:54,228] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), pow), ConstantVariable(int)] | |
[2023-07-06 04:20:54,229] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR mean [TensorVariable()] | |
[2023-07-06 04:20:54,229] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [GetAttrVariable(TensorVariable(), mean)] | |
[2023-07-06 04:20:54,229] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST True [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int)] | |
[2023-07-06 04:20:54,229] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('keepdim',) [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int), ConstantVariable(bool)] | |
[2023-07-06 04:20:54,229] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 2 [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int), ConstantVariable(bool), ConstantVariable(tuple)] | |
[2023-07-06 04:20:54,230] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST variance [TensorVariable()] | |
[2023-07-06 04:20:54,230] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86 | |
[2023-07-06 04:20:54,230] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:54,230] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [TensorVariable()] | |
[2023-07-06 04:20:54,230] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR rsqrt [TensorVariable(), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:54,231] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST variance [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:54,231] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:54,231] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR variance_epsilon [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:54,231] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(float)] | |
[2023-07-06 04:20:54,232] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:54,232] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:54,233] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:54,233] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88 | |
[2023-07-06 04:20:54,233] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,233] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR weight [NNModuleVariable()] | |
[2023-07-06 04:20:54,234] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['self'].weight (16,) [<DimDynamic.STATIC: 2>] [None] | |
[2023-07-06 04:20:54,235] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:54,235] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:54,236] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable()] | |
[2023-07-06 04:20:54,236] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST input_dtype [GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:54,236] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16)] | |
[2023-07-06 04:20:54,237] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:54,237] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo done tracing forward (RETURN_VALUE) | |
[2023-07-06 04:20:54,238] torch._dynamo.symbolic_convert: [DEBUG] RETURN_VALUE triggered compile | |
[2023-07-06 04:20:54,238] torch._dynamo.output_graph: [DEBUG] COMPILING GRAPH due to GraphCompileReason(reason='return_value', user_stack=[<FrameSummary file /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py, line 88 in forward>], graph_break=False) | |
[2023-07-06 04:20:54,238] torch._dynamo.output_graph.__graph_code: [DEBUG] TRACED GRAPH | |
===== __compiled_fn_9 ===== | |
<eval_with_key>.161 class GraphModule(torch.nn.Module): | |
def forward(self, L_hidden_states_ : torch.Tensor): | |
l_hidden_states_ = L_hidden_states_ | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85, code: variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) | |
to = l_hidden_states_.to(torch.float32) | |
pow_1 = to.pow(2); to = None | |
mean = pow_1.mean(-1, keepdim = True); pow_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86, code: hidden_states = hidden_states * torch.rsqrt(variance + self.variance_epsilon) | |
add = mean + 1e-06; mean = None | |
rsqrt = torch.rsqrt(add); add = None | |
mul = l_hidden_states_ * rsqrt; l_hidden_states_ = rsqrt = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88, code: return (self.weight * hidden_states).to(input_dtype) | |
l__self___weight = self.L__self___weight | |
mul_1 = l__self___weight * mul; l__self___weight = mul = None | |
to_1 = mul_1.to(torch.float16); mul_1 = None | |
return (to_1,) | |
[2023-07-06 04:20:54,239] torch._dynamo.output_graph.__graph: [DEBUG] TRACED GRAPH | |
__compiled_fn_9 <eval_with_key>.161 opcode name target args kwargs | |
------------- ---------------- -------------------------------------------------------- --------------------------------- ----------------- | |
placeholder l_hidden_states_ L_hidden_states_ () {} | |
call_method to to (l_hidden_states_, torch.float32) {} | |
call_method pow_1 pow (to, 2) {} | |
call_method mean mean (pow_1, -1) {'keepdim': True} | |
call_function add <built-in function add> (mean, 1e-06) {} | |
call_function rsqrt <built-in method rsqrt of type object at 0x7fb940469580> (add,) {} | |
call_function mul <built-in function mul> (l_hidden_states_, rsqrt) {} | |
get_attr l__self___weight L__self___weight () {} | |
call_function mul_1 <built-in function mul> (l__self___weight, mul) {} | |
call_method to_1 to (mul_1, torch.float16) {} | |
output output output ((to_1,),) {} | |
[2023-07-06 04:20:54,239] torch._dynamo.output_graph: [INFO] Step 2: calling compiler function inductor | |
[2023-07-06 04:20:54,258] torch._functorch.aot_autograd.__aot_graphs: [INFO] TRACED GRAPH | |
===== Forward graph 19 ===== | |
<eval_with_key>.165 class <lambda>(torch.nn.Module): | |
def forward(self, arg0_1: f16[16], arg1_1: f16[1, 1000, 16]): | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85, code: variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) | |
convert_element_type: f32[1, 1000, 16] = torch.ops.prims.convert_element_type.default(arg1_1, torch.float32) | |
pow_1: f32[1, 1000, 16] = torch.ops.aten.pow.Tensor_Scalar(convert_element_type, 2); convert_element_type = None | |
mean: f32[1, 1000, 1] = torch.ops.aten.mean.dim(pow_1, [-1], True); pow_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86, code: hidden_states = hidden_states * torch.rsqrt(variance + self.variance_epsilon) | |
add: f32[1, 1000, 1] = torch.ops.aten.add.Tensor(mean, 1e-06); mean = None | |
rsqrt: f32[1, 1000, 1] = torch.ops.aten.rsqrt.default(add); add = None | |
mul: f32[1, 1000, 16] = torch.ops.aten.mul.Tensor(arg1_1, rsqrt); arg1_1 = rsqrt = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88, code: return (self.weight * hidden_states).to(input_dtype) | |
mul_1: f32[1, 1000, 16] = torch.ops.aten.mul.Tensor(arg0_1, mul); arg0_1 = mul = None | |
convert_element_type_1: f16[1, 1000, 16] = torch.ops.prims.convert_element_type.default(mul_1, torch.float16); mul_1 = None | |
return (convert_element_type_1,) | |
[2023-07-06 04:20:54,259] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 6 | |
[2023-07-06 04:20:54,276] torch._inductor.graph: [DEBUG] Force channels last inputs for 0 conv for the current graph with id 6 | |
[2023-07-06 04:20:54,291] torch._inductor.scheduler: [DEBUG] remove_buffer('buf0') | |
[2023-07-06 04:20:54,292] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_ubuntu/tn/ctnjyv37bz6fvtmwelga5764xq272722mvbuxkb4qqgd3syt7hwl.py | |
[2023-07-06 04:20:54,293] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 6 | |
[2023-07-06 04:20:54,295] torch._dynamo.output_graph: [INFO] Step 2: done compiler function inductor | |
[2023-07-06 04:20:54,298] torch._dynamo.convert_frame.__guards: [DEBUG] GUARDS: | |
___check_obj_id(L['self'], 140431662764176) | |
___check_type_id(L['hidden_states'], 77765680) | |
hasattr(L['hidden_states'], '_dynamo_dynamic_indices') == False | |
not ___is_grad_enabled() | |
not ___are_deterministic_algorithms_enabled() | |
utils_device.CURRENT_DEVICE == None | |
str(G['torch'].float32) == 'torch.float32' | |
[2023-07-06 04:20:54,298] torch._inductor.cudagraph_trees: [DEBUG] Running warmup of function 4 | |
[2023-07-06 04:20:54,299] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo start tracing __init__ <string>:2 | |
[2023-07-06 04:20:54,299] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['last_hidden_state'] (1, 1000, 16) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None] | |
[2023-07-06 04:20:54,300] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line <string>:3 | |
[2023-07-06 04:20:54,300] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST last_hidden_state [] | |
[2023-07-06 04:20:54,300] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [TensorVariable()] | |
[2023-07-06 04:20:54,300] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_ATTR last_hidden_state [TensorVariable(), DataClassVariable()] | |
[2023-07-06 04:20:54,301] torch._dynamo.symbolic_convert: [DEBUG] STORE_ATTR triggered compile | |
Traceback (most recent call last): | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 1195, in STORE_ATTR | |
.call_function(self, [obj, ConstantVariable(inst.argval), val], {}) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/builtin.py", line 591, in call_function | |
result = handler(tx, *args, **kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/builtin.py", line 1103, in call_setattr | |
return obj.call_method(tx, "__setattr__", [name_var, val], {}) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/dicts.py", line 417, in call_method | |
return super().call_method(tx, name, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/dicts.py", line 198, in call_method | |
return super().call_method(tx, name, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/base.py", line 260, in call_method | |
raise unimplemented(f"call_method {self} {name} {args} {kwargs}") | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/exc.py", line 140, in unimplemented | |
raise Unsupported(msg) | |
torch._dynamo.exc.Unsupported: call_method DataClassVariable() __setitem__ [ConstantVariable(str), TensorVariable()] {} | |
[2023-07-06 04:20:54,304] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:54,304] torch._dynamo.output_graph: [DEBUG] COMPILING GRAPH due to GraphCompileReason(reason='store_attr', user_stack=[<FrameSummary file <string>, line 3 in __init__>], graph_break=True) | |
[2023-07-06 04:20:54,309] torch._dynamo.convert_frame.__guards: [DEBUG] GUARDS: | |
___check_type_id(L['self'], 110084784) | |
___check_type_id(L['past_key_values'], 7638272) | |
len(L['past_key_values']) == 2 | |
___check_obj_id(L['self'].attentions, 7649088) | |
hasattr(L['last_hidden_state'], '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['self'].hidden_states, 7649088) | |
___check_obj_id(L['self'].past_key_values, 7649088) | |
___check_obj_id(L['self'].last_hidden_state, 7649088) | |
not ___is_grad_enabled() | |
not ___are_deterministic_algorithms_enabled() | |
utils_device.CURRENT_DEVICE == None | |
[2023-07-06 04:20:54,309] torch._dynamo.convert_frame: [DEBUG] skipping because no torch.* <resume in __init__> <string> 3 | |
[2023-07-06 04:20:54,309] torch._dynamo.convert_frame: [DEBUG] skipping because no torch.* __post_init__ /home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/utils/generic.py 251 | |
[2023-07-06 04:20:54,309] torch._dynamo.eval_frame: [DEBUG] skipping fields /home/ubuntu/miniconda3/lib/python3.10/dataclasses.py | |
[2023-07-06 04:20:54,309] torch._dynamo.eval_frame: [DEBUG] skipping <genexpr> /home/ubuntu/miniconda3/lib/python3.10/dataclasses.py | |
[2023-07-06 04:20:54,310] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo start tracing __setitem__ /home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/utils/generic.py:328 | |
[2023-07-06 04:20:54,311] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['self'].last_hidden_state (1, 1000, 16) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None] | |
[2023-07-06 04:20:54,312] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/utils/generic.py:330 | |
[2023-07-06 04:20:54,312] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL super [] | |
[2023-07-06 04:20:54,312] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_DEREF __class__ [BuiltinVariable(super)] | |
[2023-07-06 04:20:54,312] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [BuiltinVariable(super), UserDefinedClassVariable()] | |
[2023-07-06 04:20:54,312] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 2 [BuiltinVariable(super), UserDefinedClassVariable(), DataClassVariable()] | |
[2023-07-06 04:20:54,313] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR __setitem__ [SuperVariable()] | |
[2023-07-06 04:20:54,313] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key [GetAttrVariable(SuperVariable(), __setitem__)] | |
[2023-07-06 04:20:54,313] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST value [GetAttrVariable(SuperVariable(), __setitem__), ConstantVariable(str)] | |
[2023-07-06 04:20:54,313] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 2 [GetAttrVariable(SuperVariable(), __setitem__), ConstantVariable(str), TensorVariable()] | |
[2023-07-06 04:20:54,313] torch._dynamo.symbolic_convert: [DEBUG] break_graph_if_unsupported triggered compile | |
Traceback (most recent call last): | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 390, in wrapper | |
return inner_fn(self, inst) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 1100, in CALL_FUNCTION | |
self.call_function(fn, args, {}) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 559, in call_function | |
self.push(fn.call_function(self, args, kwargs)) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/misc.py", line 588, in call_function | |
return self.obj.call_method(tx, self.name, args, kwargs).add_options(self) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/misc.py", line 115, in call_method | |
unimplemented(f"non-function or method super: {inner_fn}") | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/exc.py", line 140, in unimplemented | |
raise Unsupported(msg) | |
torch._dynamo.exc.Unsupported: non-function or method super: <slot wrapper '__setitem__' of 'collections.OrderedDict' objects> | |
[2023-07-06 04:20:54,315] torch._dynamo.symbolic_convert.__graph_breaks: [DEBUG] Graph break: non-function or method super: <slot wrapper '__setitem__' of 'collections.OrderedDict' objects> from user code at File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/utils/generic.py", line 330, in __setitem__ | |
super().__setitem__(key, value) | |
[2023-07-06 04:20:54,315] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:54,315] torch._dynamo.output_graph: [DEBUG] COMPILING GRAPH due to GraphCompileReason(reason="non-function or method super: <slot wrapper '__setitem__' of 'collections.OrderedDict' objects>", user_stack=[<FrameSummary file /home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/utils/generic.py, line 330 in __setitem__>], graph_break=True) | |
[2023-07-06 04:20:54,321] torch._dynamo.convert_frame.__guards: [DEBUG] GUARDS: | |
___check_type_id(L['key'], 7626144) | |
L['key'] == 'last_hidden_state' | |
___check_type_id(L['self'], 110084784) | |
L['self'].last_hidden_state is L['value'] | |
___check_obj_id(L['__class__'], 107106000) | |
___check_obj_id(L['self'].attentions, 7649088) | |
___check_obj_id(L['self'].hidden_states, 7649088) | |
___check_type_id(L['self'].past_key_values, 7638272) | |
len(L['self'].past_key_values) == 2 | |
hasattr(L['self'].last_hidden_state, '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['self'].past_key_values[0], 7649088) | |
___check_obj_id(L['self'].past_key_values[1], 7649088) | |
not ___is_grad_enabled() | |
not ___are_deterministic_algorithms_enabled() | |
utils_device.CURRENT_DEVICE == None | |
[2023-07-06 04:20:54,321] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo start tracing <resume in __setitem__> /home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/utils/generic.py:330 | |
[2023-07-06 04:20:54,322] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['self'].last_hidden_state (1, 1000, 16) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None] | |
[2023-07-06 04:20:54,323] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST ___stack0 [] | |
[2023-07-06 04:20:54,323] torch._dynamo.symbolic_convert: [DEBUG] TRACE JUMP_ABSOLUTE 20 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:54,323] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_TOP None [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:54,323] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/utils/generic.py:332 | |
[2023-07-06 04:20:54,323] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL super [] | |
[2023-07-06 04:20:54,323] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_DEREF __class__ [BuiltinVariable(super)] | |
[2023-07-06 04:20:54,323] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [BuiltinVariable(super), UserDefinedClassVariable()] | |
[2023-07-06 04:20:54,323] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 2 [BuiltinVariable(super), UserDefinedClassVariable(), DataClassVariable()] | |
[2023-07-06 04:20:54,323] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR __setattr__ [SuperVariable()] | |
[2023-07-06 04:20:54,324] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key [GetAttrVariable(SuperVariable(), __setattr__)] | |
[2023-07-06 04:20:54,324] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST value [GetAttrVariable(SuperVariable(), __setattr__), ConstantVariable(str)] | |
[2023-07-06 04:20:54,324] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 2 [GetAttrVariable(SuperVariable(), __setattr__), ConstantVariable(str), TensorVariable()] | |
[2023-07-06 04:20:54,324] torch._dynamo.symbolic_convert: [DEBUG] break_graph_if_unsupported triggered compile | |
Traceback (most recent call last): | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 390, in wrapper | |
return inner_fn(self, inst) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 1100, in CALL_FUNCTION | |
self.call_function(fn, args, {}) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 559, in call_function | |
self.push(fn.call_function(self, args, kwargs)) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/misc.py", line 588, in call_function | |
return self.obj.call_method(tx, self.name, args, kwargs).add_options(self) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/misc.py", line 115, in call_method | |
unimplemented(f"non-function or method super: {inner_fn}") | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/exc.py", line 140, in unimplemented | |
raise Unsupported(msg) | |
torch._dynamo.exc.Unsupported: non-function or method super: <slot wrapper '__setattr__' of 'object' objects> | |
[2023-07-06 04:20:54,324] torch._dynamo.symbolic_convert.__graph_breaks: [DEBUG] Graph break: non-function or method super: <slot wrapper '__setattr__' of 'object' objects> from user code at File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/utils/generic.py", line 332, in <resume in __setitem__> | |
super().__setattr__(key, value) | |
[2023-07-06 04:20:54,324] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:54,324] torch._dynamo.output_graph: [DEBUG] COMPILING GRAPH due to GraphCompileReason(reason="non-function or method super: <slot wrapper '__setattr__' of 'object' objects>", user_stack=[<FrameSummary file /home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/utils/generic.py, line 332 in <resume in __setitem__>>], graph_break=True) | |
[2023-07-06 04:20:54,330] torch._dynamo.convert_frame.__guards: [DEBUG] GUARDS: | |
___check_type_id(L['key'], 7626144) | |
L['key'] == 'last_hidden_state' | |
___check_type_id(L['self'], 110084784) | |
L['self'].last_hidden_state is L['value'] | |
___check_obj_id(L['__class__'], 107106000) | |
___check_obj_id(L['self'].attentions, 7649088) | |
___check_obj_id(L['self'].hidden_states, 7649088) | |
___check_type_id(L['self'].past_key_values, 7638272) | |
len(L['self'].past_key_values) == 2 | |
hasattr(L['self'].last_hidden_state, '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['self'].past_key_values[0], 7649088) | |
___check_obj_id(L['self'].past_key_values[1], 7649088) | |
not ___is_grad_enabled() | |
not ___are_deterministic_algorithms_enabled() | |
utils_device.CURRENT_DEVICE == None | |
[2023-07-06 04:20:54,330] torch._dynamo.convert_frame: [DEBUG] skipping because no torch.* <resume in __setitem__> /home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/utils/generic.py 332 | |
[2023-07-06 04:20:54,330] torch._dynamo.convert_frame.__recompiles: [DEBUG] ('Recompiling function __setitem__ in /home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/utils/generic.py', 'set env var TORCHDYNAMO_REPORT_GUARD_FAILURES=1 to debug further') | |
[2023-07-06 04:20:54,330] torch._dynamo.convert_frame: [DEBUG] skipping because no torch.* __setitem__ /home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/utils/generic.py 328 | |
[2023-07-06 04:20:54,331] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo start tracing <resume in forward> /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:692 | |
[2023-07-06 04:20:54,332] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['___stack0'].last_hidden_state (1, 1000, 16) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None] | |
[2023-07-06 04:20:54,333] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST ___stack0 [] | |
[2023-07-06 04:20:54,333] torch._dynamo.symbolic_convert: [DEBUG] TRACE JUMP_ABSOLUTE 90 [DataClassVariable()] | |
[2023-07-06 04:20:54,333] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST outputs [DataClassVariable()] | |
[2023-07-06 04:20:54,333] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:704 | |
[2023-07-06 04:20:54,333] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST outputs [] | |
[2023-07-06 04:20:54,333] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [DataClassVariable()] | |
[2023-07-06 04:20:54,333] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [DataClassVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:54,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:54,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:705 | |
[2023-07-06 04:20:54,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR lm_head [NNModuleVariable()] | |
[2023-07-06 04:20:54,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [NNModuleVariable()] | |
[2023-07-06 04:20:54,334] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:54,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST logits [TensorVariable()] | |
[2023-07-06 04:20:54,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:707 | |
[2023-07-06 04:20:54,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [] | |
[2023-07-06 04:20:54,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST loss [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:54,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:708 | |
[2023-07-06 04:20:54,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST labels [] | |
[2023-07-06 04:20:54,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:54,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE IS_OP 1 [ConstantVariable(NoneType), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:54,337] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 222 [ConstantVariable(bool)] | |
[2023-07-06 04:20:54,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:720 | |
[2023-07-06 04:20:54,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST return_dict [] | |
[2023-07-06 04:20:54,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 266 [ConstantVariable(bool)] | |
[2023-07-06 04:20:54,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:724 | |
[2023-07-06 04:20:54,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL CausalLMOutputWithPast [] | |
[2023-07-06 04:20:54,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:725 | |
[2023-07-06 04:20:54,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST loss [UserDefinedClassVariable()] | |
[2023-07-06 04:20:54,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:726 | |
[2023-07-06 04:20:54,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST logits [UserDefinedClassVariable(), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:54,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:727 | |
[2023-07-06 04:20:54,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST outputs [UserDefinedClassVariable(), ConstantVariable(NoneType), TensorVariable()] | |
[2023-07-06 04:20:54,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR past_key_values [UserDefinedClassVariable(), ConstantVariable(NoneType), TensorVariable(), DataClassVariable()] | |
[2023-07-06 04:20:54,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:728 | |
[2023-07-06 04:20:54,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST outputs [UserDefinedClassVariable(), ConstantVariable(NoneType), TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:54,338] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR hidden_states [UserDefinedClassVariable(), ConstantVariable(NoneType), TensorVariable(), TupleVariable(), DataClassVariable()] | |
[2023-07-06 04:20:54,339] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:729 | |
[2023-07-06 04:20:54,339] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST outputs [UserDefinedClassVariable(), ConstantVariable(NoneType), TensorVariable(), TupleVariable(), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:54,339] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR attentions [UserDefinedClassVariable(), ConstantVariable(NoneType), TensorVariable(), TupleVariable(), ConstantVariable(NoneType), DataClassVariable()] | |
[2023-07-06 04:20:54,339] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:724 | |
[2023-07-06 04:20:54,339] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('loss', 'logits', 'past_key_values', 'hidden_states', 'attentions') [UserDefinedClassVariable(), ConstantVariable(NoneType), TensorVariable(), TupleVariable(), ConstantVariable(NoneType), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:54,339] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 5 [UserDefinedClassVariable(), ConstantVariable(NoneType), TensorVariable(), TupleVariable(), ConstantVariable(NoneType), ConstantVariable(NoneType), ConstantVariable(tuple)] | |
[2023-07-06 04:20:54,339] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [DataClassVariable()] | |
[2023-07-06 04:20:54,339] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo done tracing <resume in forward> (RETURN_VALUE) | |
[2023-07-06 04:20:54,339] torch._dynamo.symbolic_convert: [DEBUG] RETURN_VALUE triggered compile | |
[2023-07-06 04:20:54,340] torch._dynamo.output_graph: [DEBUG] COMPILING GRAPH due to GraphCompileReason(reason='return_value', user_stack=[<FrameSummary file /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py, line 724 in <resume in forward>>], graph_break=False) | |
[2023-07-06 04:20:54,341] torch._dynamo.output_graph.__graph_code: [DEBUG] TRACED GRAPH | |
===== __compiled_fn_13 ===== | |
<eval_with_key>.167 class GraphModule(torch.nn.Module): | |
def forward(self, L_stack0_last_hidden_state : torch.Tensor): | |
l_stack0_last_hidden_state = L_stack0_last_hidden_state | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:705, code: logits = self.lm_head(hidden_states) | |
l__self___lm_head = self.L__self___lm_head(l_stack0_last_hidden_state); l_stack0_last_hidden_state = None | |
return (l__self___lm_head,) | |
[2023-07-06 04:20:54,341] torch._dynamo.output_graph.__graph: [DEBUG] TRACED GRAPH | |
__compiled_fn_13 <eval_with_key>.167 opcode name target args kwargs | |
----------- -------------------------- -------------------------- ----------------------------- -------- | |
placeholder l_stack0_last_hidden_state L_stack0_last_hidden_state () {} | |
call_module l__self___lm_head L__self___lm_head (l_stack0_last_hidden_state,) {} | |
output output output ((l__self___lm_head,),) {} | |
[2023-07-06 04:20:54,341] torch._dynamo.output_graph: [INFO] Step 2: calling compiler function inductor | |
[2023-07-06 04:20:54,352] torch._functorch.aot_autograd.__aot_graphs: [INFO] TRACED GRAPH | |
===== Forward graph 20 ===== | |
<eval_with_key>.171 class <lambda>(torch.nn.Module): | |
def forward(self, arg0_1: f16[32000, 16], arg1_1: f16[1, 1000, 16]): | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:705, code: logits = self.lm_head(hidden_states) | |
permute: f16[16, 32000] = torch.ops.aten.permute.default(arg0_1, [1, 0]); arg0_1 = None | |
view: f16[1000, 16] = torch.ops.aten.view.default(arg1_1, [1000, 16]); arg1_1 = None | |
mm: f16[1000, 32000] = torch.ops.aten.mm.default(view, permute); view = permute = None | |
view_1: f16[1, 1000, 32000] = torch.ops.aten.view.default(mm, [1, 1000, 32000]); mm = None | |
return (view_1,) | |
[2023-07-06 04:20:54,361] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 7 | |
[2023-07-06 04:20:54,367] torch._inductor.graph: [DEBUG] Force channels last inputs for 0 conv for the current graph with id 7 | |
[2023-07-06 04:20:54,370] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_ubuntu/kk/ckkkx2tf6bgblkt5hisgibketyheywf6ccyhzdrdgv6pbbtyn5rf.py | |
[2023-07-06 04:20:54,370] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 7 | |
[2023-07-06 04:20:54,372] torch._dynamo.output_graph: [INFO] Step 2: done compiler function inductor | |
[2023-07-06 04:20:54,377] torch._dynamo.convert_frame.__guards: [DEBUG] GUARDS: | |
___check_obj_id(L['self'], 140431662761824) | |
___check_obj_id(L['labels'], 7649088) | |
___check_type_id(L['___stack0'], 110084784) | |
___check_obj_id(L['return_dict'], 7698144) | |
___check_obj_id(L['___stack0'].attentions, 7649088) | |
___check_obj_id(L['___stack0'].hidden_states, 7649088) | |
___check_type_id(L['___stack0'].past_key_values, 7638272) | |
len(L['___stack0'].past_key_values) == 2 | |
hasattr(L['___stack0'].last_hidden_state, '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['___stack0'].past_key_values[0], 7649088) | |
___check_obj_id(L['___stack0'].past_key_values[1], 7649088) | |
not ___is_grad_enabled() | |
not ___are_deterministic_algorithms_enabled() | |
utils_device.CURRENT_DEVICE == None | |
[2023-07-06 04:20:54,378] torch._inductor.cudagraph_trees: [DEBUG] Running warmup of function 5 | |
[2023-07-06 04:20:54,418] torch._dynamo.convert_frame.__recompiles: [DEBUG] ('Recompiling function forward in /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py', 'set env var TORCHDYNAMO_REPORT_GUARD_FAILURES=1 to debug further') | |
[2023-07-06 04:20:54,419] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo start tracing forward /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:644 | |
[2023-07-06 04:20:54,420] torch._dynamo.variables.builder: [DEBUG] automatic dynamic L['input_ids'] size(1) 1 != 1000 | |
[2023-07-06 04:20:54,420] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['input_ids'] (1, 1) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None] | |
[2023-07-06 04:20:54,421] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['attention_mask'] (1, 1003) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None] | |
[2023-07-06 04:20:54,422] torch._dynamo.variables.builder: [DEBUG] automatic dynamic L['position_ids'] size(1) 1 != 1000 | |
[2023-07-06 04:20:54,422] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['position_ids'] (1, 1) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None] | |
[2023-07-06 04:20:54,422] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['past_key_values'][0] (2, 1, 4, 1003, 4) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None, None, None] | |
[2023-07-06 04:20:54,423] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['past_key_values'][1] (2, 1, 4, 1003, 4) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None, None, None] | |
[2023-07-06 04:20:54,424] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['valid_past_index'] () [] [] | |
[2023-07-06 04:20:54,424] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:685 | |
[2023-07-06 04:20:54,424] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST output_attentions [] | |
[2023-07-06 04:20:54,424] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:54,424] torch._dynamo.symbolic_convert: [DEBUG] TRACE IS_OP 1 [ConstantVariable(NoneType), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:54,424] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 12 [ConstantVariable(bool)] | |
[2023-07-06 04:20:54,425] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,425] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR config [NNModuleVariable()] | |
[2023-07-06 04:20:54,425] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR output_attentions [HFPretrainedConfigVariable()] | |
[2023-07-06 04:20:54,425] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST output_attentions [ConstantVariable(bool)] | |
[2023-07-06 04:20:54,425] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:687 | |
[2023-07-06 04:20:54,425] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST output_hidden_states [] | |
[2023-07-06 04:20:54,425] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:54,425] torch._dynamo.symbolic_convert: [DEBUG] TRACE IS_OP 1 [ConstantVariable(NoneType), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:54,426] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 32 [ConstantVariable(bool)] | |
[2023-07-06 04:20:54,426] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,426] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR config [NNModuleVariable()] | |
[2023-07-06 04:20:54,426] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR output_hidden_states [HFPretrainedConfigVariable()] | |
[2023-07-06 04:20:54,426] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:686 | |
[2023-07-06 04:20:54,426] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST output_hidden_states [ConstantVariable(bool)] | |
[2023-07-06 04:20:54,426] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:689 | |
[2023-07-06 04:20:54,426] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST return_dict [] | |
[2023-07-06 04:20:54,426] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [ConstantVariable(bool)] | |
[2023-07-06 04:20:54,426] torch._dynamo.symbolic_convert: [DEBUG] TRACE IS_OP 1 [ConstantVariable(bool), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:54,426] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 52 [ConstantVariable(bool)] | |
[2023-07-06 04:20:54,427] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST return_dict [] | |
[2023-07-06 04:20:54,427] torch._dynamo.symbolic_convert: [DEBUG] TRACE JUMP_FORWARD 58 [ConstantVariable(bool)] | |
[2023-07-06 04:20:54,427] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST return_dict [ConstantVariable(bool)] | |
[2023-07-06 04:20:54,427] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:692 | |
[2023-07-06 04:20:54,427] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,427] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR model [NNModuleVariable()] | |
[2023-07-06 04:20:54,428] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:693 | |
[2023-07-06 04:20:54,428] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST input_ids [NNModuleVariable()] | |
[2023-07-06 04:20:54,428] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:694 | |
[2023-07-06 04:20:54,428] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST attention_mask [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:54,428] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:695 | |
[2023-07-06 04:20:54,428] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST position_ids [NNModuleVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:54,428] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:696 | |
[2023-07-06 04:20:54,428] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST past_key_values [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:54,428] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:697 | |
[2023-07-06 04:20:54,428] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST inputs_embeds [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ListVariable()] | |
[2023-07-06 04:20:54,428] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:698 | |
[2023-07-06 04:20:54,429] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST output_attentions [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ListVariable(), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:54,429] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:699 | |
[2023-07-06 04:20:54,429] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST output_hidden_states [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ListVariable(), ConstantVariable(NoneType), ConstantVariable(bool)] | |
[2023-07-06 04:20:54,429] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:700 | |
[2023-07-06 04:20:54,429] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST return_dict [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ListVariable(), ConstantVariable(NoneType), ConstantVariable(bool), ConstantVariable(bool)] | |
[2023-07-06 04:20:54,429] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:701 | |
[2023-07-06 04:20:54,429] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ListVariable(), ConstantVariable(NoneType), ConstantVariable(bool), ConstantVariable(bool), ConstantVariable(bool)] | |
[2023-07-06 04:20:54,429] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:692 | |
[2023-07-06 04:20:54,429] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('input_ids', 'attention_mask', 'position_ids', 'past_key_values', 'inputs_embeds', 'output_attentions', 'output_hidden_states', 'return_dict', 'valid_past_index') [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ListVariable(), ConstantVariable(NoneType), ConstantVariable(bool), ConstantVariable(bool), ConstantVariable(bool), TensorVariable()] | |
[2023-07-06 04:20:54,429] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 9 [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ListVariable(), ConstantVariable(NoneType), ConstantVariable(bool), ConstantVariable(bool), ConstantVariable(bool), TensorVariable(), ConstantVariable(tuple)] | |
[2023-07-06 04:20:54,429] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:54,536] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:54,536] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:54,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:54,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:54,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:54,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:54,538] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,538] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:54,538] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function LlamaModel.forward at 0x7fb8d0c46cb0>, NNModuleVariable())] | |
[2023-07-06 04:20:54,538] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,538] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,538] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:54,538] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,539] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,539] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:54,539] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,539] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,539] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:54,539] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,539] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,539] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:54,540] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,540] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:54,540] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:54,540] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,540] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,540] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:54,540] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:54,540] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,540] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,540] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:54,540] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:54,540] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,540] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,540] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:54,540] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:54,541] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,541] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,541] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:54,541] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:54,541] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function LlamaModel.forward at 0x7fb8d0c46cb0>, NNModuleVariable())] | |
[2023-07-06 04:20:54,541] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function LlamaModel.forward at 0x7fb8d0c46cb0>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:54,541] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function LlamaModel.forward at 0x7fb8d0c46cb0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:54,541] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function LlamaModel.forward at 0x7fb8d0c46cb0>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:54,541] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function LlamaModel.forward at 0x7fb8d0c46cb0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:54,542] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:54,542] torch._dynamo.symbolic_convert: [DEBUG] FAILED INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:54,542] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:54,542] torch._dynamo.symbolic_convert: [DEBUG] break_graph_if_unsupported triggered compile | |
Traceback (most recent call last): | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2187, in inline_call_ | |
sub_locals, closure_cells = func.bind_args(parent, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 187, in bind_args | |
[ | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 188, in <listcomp> | |
wrap(val=arg, source=source) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 62, in wrap_bound_arg | |
return VariableBuilder(tx, source=source, **options)(val) | |
TypeError: VariableBuilder.__init__() got an unexpected keyword argument 'guards' | |
During handling of the above exception, another exception occurred: | |
Traceback (most recent call last): | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 390, in wrapper | |
return inner_fn(self, inst) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 1152, in CALL_FUNCTION_KW | |
self.call_function(fn, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 559, in call_function | |
self.push(fn.call_function(self, args, kwargs)) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/nn_module.py", line 333, in call_function | |
return tx.inline_user_function_return( | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 595, in inline_user_function_return | |
result = InliningInstructionTranslator.inline_call(self, fn, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2132, in inline_call | |
return cls.inline_call_(parent, func, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2229, in inline_call_ | |
tracer.run() | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 708, in run | |
and self.step() | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 668, in step | |
getattr(self, inst.opname)(inst) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 390, in wrapper | |
return inner_fn(self, inst) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 1140, in CALL_FUNCTION_EX | |
self.call_function(fn, argsvars.items, kwargsvars.items) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 559, in call_function | |
self.push(fn.call_function(self, args, kwargs)) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 333, in call_function | |
return super().call_function(tx, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 287, in call_function | |
return super().call_function(tx, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 120, in call_function | |
return tx.inline_user_function_return( | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 595, in inline_user_function_return | |
result = InliningInstructionTranslator.inline_call(self, fn, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2132, in inline_call | |
return cls.inline_call_(parent, func, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2190, in inline_call_ | |
raise ArgsMismatchError( | |
torch._dynamo.exc.ArgsMismatchError: VariableBuilder.__init__() got an unexpected keyword argument 'guards'. | |
func = 'forward' /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:477, args = [<class 'trfs_fast.llama.LlamaModel'>], kwargs = OrderedDict([('input_ids', TensorVariable()), ('attention_mask', TensorVariable()), ('position_ids', TensorVariable()), ('past_key_values', ListVariable()), ('inputs_embeds', ConstantVariable(NoneType)), ('output_attentions', ConstantVariable(bool)), ('output_hidden_states', ConstantVariable(bool)), ('return_dict', ConstantVariable(bool)), ('valid_past_index', TensorVariable())]) | |
[2023-07-06 04:20:54,542] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:54,542] torch._dynamo.output_graph: [DEBUG] COMPILING GRAPH due to GraphCompileReason(reason="VariableBuilder.__init__() got an unexpected keyword argument 'guards'.\n func = 'forward' /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:477, args = [<class 'trfs_fast.llama.LlamaModel'>], kwargs = OrderedDict([('input_ids', TensorVariable()), ('attention_mask', TensorVariable()), ('position_ids', TensorVariable()), ('past_key_values', ListVariable()), ('inputs_embeds', ConstantVariable(NoneType)), ('output_attentions', ConstantVariable(bool)), ('output_hidden_states', ConstantVariable(bool)), ('return_dict', ConstantVariable(bool)), ('valid_past_index', TensorVariable())])", user_stack=[<FrameSummary file /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py, line 692 in forward>, <FrameSummary file /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py, line 1514 in _call_impl>], graph_break=True) | |
[2023-07-06 04:20:54,551] torch._dynamo.convert_frame.__guards: [DEBUG] GUARDS: | |
___check_obj_id(L['self'], 140431662761824) | |
hasattr(L['input_ids'], '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['return_dict'], 7698144) | |
hasattr(L['position_ids'], '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['inputs_embeds'], 7649088) | |
hasattr(L['attention_mask'], '_dynamo_dynamic_indices') == False | |
___check_type_id(L['past_key_values'], 7662688) | |
len(L['past_key_values']) == 2 | |
hasattr(L['valid_past_index'], '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['output_attentions'], 7649088) | |
hasattr(L['past_key_values'][0], '_dynamo_dynamic_indices') == False | |
hasattr(L['past_key_values'][1], '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['output_hidden_states'], 7649088) | |
not ___is_grad_enabled() | |
not ___are_deterministic_algorithms_enabled() | |
utils_device.CURRENT_DEVICE == None | |
[2023-07-06 04:20:54,551] torch._dynamo.convert_frame.__recompiles: [DEBUG] ('Recompiling function forward in /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py', 'set env var TORCHDYNAMO_REPORT_GUARD_FAILURES=1 to debug further') | |
[2023-07-06 04:20:54,552] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo start tracing forward /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:261 | |
[2023-07-06 04:20:54,552] torch._dynamo.variables.builder: [DEBUG] automatic dynamic L['hidden_states'] size(1) 1 != 1000 | |
[2023-07-06 04:20:54,552] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['hidden_states'] (1, 1, 16) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None] | |
[2023-07-06 04:20:54,553] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['attention_mask'] (1, 1003) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None] | |
[2023-07-06 04:20:54,554] torch._dynamo.variables.builder: [DEBUG] automatic dynamic L['position_ids'] size(1) 1 != 1000 | |
[2023-07-06 04:20:54,554] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['position_ids'] (1, 1) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None] | |
[2023-07-06 04:20:54,555] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['past_key_value'] (2, 1, 4, 1003, 4) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None, None, None] | |
[2023-07-06 04:20:54,556] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['valid_past_index'] () [] [] | |
[2023-07-06 04:20:54,556] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:281 | |
[2023-07-06 04:20:54,556] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:54,556] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST residual [TensorVariable()] | |
[2023-07-06 04:20:54,556] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:283 | |
[2023-07-06 04:20:54,556] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,556] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR input_layernorm [NNModuleVariable()] | |
[2023-07-06 04:20:54,556] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [NNModuleVariable()] | |
[2023-07-06 04:20:54,557] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:54,557] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:54,558] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:54,558] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:54,558] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:54,559] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:54,559] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:54,559] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:54,559] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,559] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:54,559] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable())] | |
[2023-07-06 04:20:54,560] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,560] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,560] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:54,560] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,560] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,560] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:54,560] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,560] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,560] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:54,561] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,561] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,561] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:54,561] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,561] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:54,561] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:54,561] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,561] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,561] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:54,561] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:54,561] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,561] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,561] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:54,562] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:54,562] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,562] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,562] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:54,562] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:54,562] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,562] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,562] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:54,562] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:54,562] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable())] | |
[2023-07-06 04:20:54,562] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:54,562] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:54,562] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:54,562] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:54,563] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object forward at 0x7fb8cddd7ec0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 83> | |
[2023-07-06 04:20:54,563] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:84 | |
[2023-07-06 04:20:54,563] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:54,563] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR dtype [TensorVariable()] | |
[2023-07-06 04:20:54,566] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST input_dtype [TorchVariable(torch.float16)] | |
[2023-07-06 04:20:54,566] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85 | |
[2023-07-06 04:20:54,566] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:54,566] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable()] | |
[2023-07-06 04:20:54,566] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:54,566] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR float32 [GetAttrVariable(TensorVariable(), to), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:54,566] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), to), ConstantVariable(dtype)] | |
[2023-07-06 04:20:54,568] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR pow [TensorVariable()] | |
[2023-07-06 04:20:54,568] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [GetAttrVariable(TensorVariable(), pow)] | |
[2023-07-06 04:20:54,568] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), pow), ConstantVariable(int)] | |
[2023-07-06 04:20:54,569] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR mean [TensorVariable()] | |
[2023-07-06 04:20:54,569] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [GetAttrVariable(TensorVariable(), mean)] | |
[2023-07-06 04:20:54,569] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST True [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int)] | |
[2023-07-06 04:20:54,569] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('keepdim',) [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int), ConstantVariable(bool)] | |
[2023-07-06 04:20:54,569] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 2 [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int), ConstantVariable(bool), ConstantVariable(tuple)] | |
[2023-07-06 04:20:54,570] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST variance [TensorVariable()] | |
[2023-07-06 04:20:54,570] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86 | |
[2023-07-06 04:20:54,570] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:54,570] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [TensorVariable()] | |
[2023-07-06 04:20:54,570] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR rsqrt [TensorVariable(), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:54,571] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST variance [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:54,571] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:54,571] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR variance_epsilon [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:54,571] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(float)] | |
[2023-07-06 04:20:54,572] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:54,572] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:54,573] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:54,573] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88 | |
[2023-07-06 04:20:54,573] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,573] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR weight [NNModuleVariable()] | |
[2023-07-06 04:20:54,574] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['self'].input_layernorm.weight (16,) [<DimDynamic.STATIC: 2>] [None] | |
[2023-07-06 04:20:54,574] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:54,574] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:54,575] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable()] | |
[2023-07-06 04:20:54,575] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST input_dtype [GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:54,575] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16)] | |
[2023-07-06 04:20:54,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:54,576] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object forward at 0x7fb8cddd7ec0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 83> | |
[2023-07-06 04:20:54,577] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:54,577] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:54,577] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:54,577] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:286 | |
[2023-07-06 04:20:54,577] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,577] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR self_attn [NNModuleVariable()] | |
[2023-07-06 04:20:54,577] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:287 | |
[2023-07-06 04:20:54,577] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [NNModuleVariable()] | |
[2023-07-06 04:20:54,577] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:288 | |
[2023-07-06 04:20:54,577] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST attention_mask [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:54,577] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:289 | |
[2023-07-06 04:20:54,577] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST position_ids [NNModuleVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:54,577] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:290 | |
[2023-07-06 04:20:54,577] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST past_key_value [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:54,577] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:291 | |
[2023-07-06 04:20:54,577] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST output_attentions [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:54,577] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:292 | |
[2023-07-06 04:20:54,577] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ConstantVariable(bool)] | |
[2023-07-06 04:20:54,578] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:286 | |
[2023-07-06 04:20:54,578] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('hidden_states', 'attention_mask', 'position_ids', 'past_key_value', 'output_attentions', 'valid_past_index') [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ConstantVariable(bool), TensorVariable()] | |
[2023-07-06 04:20:54,578] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 6 [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ConstantVariable(bool), TensorVariable(), ConstantVariable(tuple)] | |
[2023-07-06 04:20:54,578] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:54,579] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:54,579] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:54,580] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:54,580] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:54,580] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:54,580] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:54,580] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,580] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:54,581] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function LlamaAttention.forward at 0x7fb8d0c46680>, NNModuleVariable())] | |
[2023-07-06 04:20:54,581] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,581] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,581] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:54,581] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,581] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,581] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:54,581] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,581] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,581] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:54,582] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,582] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:54,582] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:54,582] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,582] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:54,582] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:54,582] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,582] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,582] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:54,582] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:54,582] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,582] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,582] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:54,582] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:54,582] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,582] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,582] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:54,582] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:54,583] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:54,583] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:54,583] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:54,583] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:54,583] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function LlamaAttention.forward at 0x7fb8d0c46680>, NNModuleVariable())] | |
[2023-07-06 04:20:54,583] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function LlamaAttention.forward at 0x7fb8d0c46680>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:54,583] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function LlamaAttention.forward at 0x7fb8d0c46680>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:54,583] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function LlamaAttention.forward at 0x7fb8d0c46680>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:54,583] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function LlamaAttention.forward at 0x7fb8d0c46680>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:54,583] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:54,583] torch._dynamo.symbolic_convert: [DEBUG] FAILED INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:54,583] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:54,584] torch._dynamo.symbolic_convert: [DEBUG] break_graph_if_unsupported triggered compile | |
Traceback (most recent call last): | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2187, in inline_call_ | |
sub_locals, closure_cells = func.bind_args(parent, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 187, in bind_args | |
[ | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 188, in <listcomp> | |
wrap(val=arg, source=source) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 62, in wrap_bound_arg | |
return VariableBuilder(tx, source=source, **options)(val) | |
TypeError: VariableBuilder.__init__() got an unexpected keyword argument 'guards' | |
During handling of the above exception, another exception occurred: | |
Traceback (most recent call last): | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 390, in wrapper | |
return inner_fn(self, inst) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 1152, in CALL_FUNCTION_KW | |
self.call_function(fn, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 559, in call_function | |
self.push(fn.call_function(self, args, kwargs)) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/nn_module.py", line 333, in call_function | |
return tx.inline_user_function_return( | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 595, in inline_user_function_return | |
result = InliningInstructionTranslator.inline_call(self, fn, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2132, in inline_call | |
return cls.inline_call_(parent, func, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2229, in inline_call_ | |
tracer.run() | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 708, in run | |
and self.step() | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 668, in step | |
getattr(self, inst.opname)(inst) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 390, in wrapper | |
return inner_fn(self, inst) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 1140, in CALL_FUNCTION_EX | |
self.call_function(fn, argsvars.items, kwargsvars.items) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 559, in call_function | |
self.push(fn.call_function(self, args, kwargs)) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 333, in call_function | |
return super().call_function(tx, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 287, in call_function | |
return super().call_function(tx, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 120, in call_function | |
return tx.inline_user_function_return( | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 595, in inline_user_function_return | |
result = InliningInstructionTranslator.inline_call(self, fn, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2132, in inline_call | |
return cls.inline_call_(parent, func, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2190, in inline_call_ | |
raise ArgsMismatchError( | |
torch._dynamo.exc.ArgsMismatchError: VariableBuilder.__init__() got an unexpected keyword argument 'guards'. | |
func = 'forward' /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:190, args = [<class 'trfs_fast.llama.LlamaAttention'>], kwargs = OrderedDict([('hidden_states', TensorVariable()), ('attention_mask', TensorVariable()), ('position_ids', TensorVariable()), ('past_key_value', TensorVariable()), ('output_attentions', ConstantVariable(bool)), ('valid_past_index', TensorVariable())]) | |
[2023-07-06 04:20:54,584] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:54,584] torch._dynamo.output_graph: [DEBUG] COMPILING GRAPH due to GraphCompileReason(reason="VariableBuilder.__init__() got an unexpected keyword argument 'guards'.\n func = 'forward' /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:190, args = [<class 'trfs_fast.llama.LlamaAttention'>], kwargs = OrderedDict([('hidden_states', TensorVariable()), ('attention_mask', TensorVariable()), ('position_ids', TensorVariable()), ('past_key_value', TensorVariable()), ('output_attentions', ConstantVariable(bool)), ('valid_past_index', TensorVariable())])", user_stack=[<FrameSummary file /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py, line 286 in forward>, <FrameSummary file /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py, line 1514 in _call_impl>], graph_break=True) | |
[2023-07-06 04:20:54,585] torch._dynamo.output_graph: [DEBUG] REMOVE UNUSED GRAPHARG L['attention_mask'] | |
[2023-07-06 04:20:54,585] torch._dynamo.output_graph: [DEBUG] REMOVE UNUSED GRAPHARG L['position_ids'] | |
[2023-07-06 04:20:54,585] torch._dynamo.output_graph: [DEBUG] REMOVE UNUSED GRAPHARG L['past_key_value'] | |
[2023-07-06 04:20:54,585] torch._dynamo.output_graph: [DEBUG] REMOVE UNUSED GRAPHARG L['valid_past_index'] | |
[2023-07-06 04:20:54,586] torch._dynamo.output_graph.__graph_code: [DEBUG] TRACED GRAPH | |
===== __compiled_fn_15 ===== | |
<eval_with_key>.176 class GraphModule(torch.nn.Module): | |
def forward(self, L_hidden_states_ : torch.Tensor): | |
l_hidden_states_ = L_hidden_states_ | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85, code: variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) | |
to = l_hidden_states_.to(torch.float32) | |
pow_1 = to.pow(2); to = None | |
mean = pow_1.mean(-1, keepdim = True); pow_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86, code: hidden_states = hidden_states * torch.rsqrt(variance + self.variance_epsilon) | |
add = mean + 1e-06; mean = None | |
rsqrt = torch.rsqrt(add); add = None | |
mul = l_hidden_states_ * rsqrt; l_hidden_states_ = rsqrt = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88, code: return (self.weight * hidden_states).to(input_dtype) | |
l__self___input_layernorm_weight = self.L__self___input_layernorm_weight | |
mul_1 = l__self___input_layernorm_weight * mul; l__self___input_layernorm_weight = mul = None | |
to_1 = mul_1.to(torch.float16); mul_1 = None | |
return (to_1,) | |
[2023-07-06 04:20:54,586] torch._dynamo.output_graph.__graph: [DEBUG] TRACED GRAPH | |
__compiled_fn_15 <eval_with_key>.176 opcode name target args kwargs | |
------------- -------------------------------- -------------------------------------------------------- --------------------------------------- ----------------- | |
placeholder l_hidden_states_ L_hidden_states_ () {} | |
call_method to to (l_hidden_states_, torch.float32) {} | |
call_method pow_1 pow (to, 2) {} | |
call_method mean mean (pow_1, -1) {'keepdim': True} | |
call_function add <built-in function add> (mean, 1e-06) {} | |
call_function rsqrt <built-in method rsqrt of type object at 0x7fb940469580> (add,) {} | |
call_function mul <built-in function mul> (l_hidden_states_, rsqrt) {} | |
get_attr l__self___input_layernorm_weight L__self___input_layernorm_weight () {} | |
call_function mul_1 <built-in function mul> (l__self___input_layernorm_weight, mul) {} | |
call_method to_1 to (mul_1, torch.float16) {} | |
output output output ((to_1,),) {} | |
[2023-07-06 04:20:54,587] torch._dynamo.output_graph: [INFO] Step 2: calling compiler function inductor | |
[2023-07-06 04:20:54,605] torch._functorch.aot_autograd.__aot_graphs: [INFO] TRACED GRAPH | |
===== Forward graph 21 ===== | |
<eval_with_key>.180 class <lambda>(torch.nn.Module): | |
def forward(self, arg0_1: f16[16], arg1_1: f16[1, 1, 16]): | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85, code: variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) | |
convert_element_type: f32[1, 1, 16] = torch.ops.prims.convert_element_type.default(arg1_1, torch.float32) | |
pow_1: f32[1, 1, 16] = torch.ops.aten.pow.Tensor_Scalar(convert_element_type, 2); convert_element_type = None | |
mean: f32[1, 1, 1] = torch.ops.aten.mean.dim(pow_1, [-1], True); pow_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86, code: hidden_states = hidden_states * torch.rsqrt(variance + self.variance_epsilon) | |
add: f32[1, 1, 1] = torch.ops.aten.add.Tensor(mean, 1e-06); mean = None | |
rsqrt: f32[1, 1, 1] = torch.ops.aten.rsqrt.default(add); add = None | |
mul: f32[1, 1, 16] = torch.ops.aten.mul.Tensor(arg1_1, rsqrt); arg1_1 = rsqrt = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88, code: return (self.weight * hidden_states).to(input_dtype) | |
mul_1: f32[1, 1, 16] = torch.ops.aten.mul.Tensor(arg0_1, mul); arg0_1 = mul = None | |
convert_element_type_1: f16[1, 1, 16] = torch.ops.prims.convert_element_type.default(mul_1, torch.float16); mul_1 = None | |
return (convert_element_type_1,) | |
[2023-07-06 04:20:54,607] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 8 | |
[2023-07-06 04:20:54,623] torch._inductor.graph: [DEBUG] Force channels last inputs for 0 conv for the current graph with id 8 | |
[2023-07-06 04:20:54,632] torch._inductor.scheduler: [DEBUG] remove_buffer('buf0') | |
[2023-07-06 04:20:55,244] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 1 configs | |
[2023-07-06 04:20:55,244] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 1, num_warps: 2, num_stages: 1 | |
[2023-07-06 04:20:55,339] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 1 configs | |
[2023-07-06 04:20:55,339] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 1, num_warps: 2, num_stages: 1 | |
[2023-07-06 04:20:55,344] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_ubuntu/vz/cvzk75oefn2tjj3ekdj3gg5a7rpdjqti5tysf4vatbmcjiobtp5i.py | |
[2023-07-06 04:20:55,344] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 8 | |
[2023-07-06 04:20:55,347] torch._dynamo.output_graph: [INFO] Step 2: done compiler function inductor | |
[2023-07-06 04:20:55,355] torch._dynamo.convert_frame.__guards: [DEBUG] GUARDS: | |
___check_obj_id(L['self'], 140431662763072) | |
hasattr(L['position_ids'], '_dynamo_dynamic_indices') == False | |
___check_type_id(L['hidden_states'], 77765680) | |
hasattr(L['hidden_states'], '_dynamo_dynamic_indices') == False | |
hasattr(L['attention_mask'], '_dynamo_dynamic_indices') == False | |
hasattr(L['past_key_value'], '_dynamo_dynamic_indices') == False | |
hasattr(L['valid_past_index'], '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['output_attentions'], 7698112) | |
not ___is_grad_enabled() | |
not ___are_deterministic_algorithms_enabled() | |
utils_device.CURRENT_DEVICE == None | |
str(G['torch'].float32) == 'torch.float32' | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks.keys()) == set() | |
[2023-07-06 04:20:55,355] torch._inductor.cudagraph_trees: [DEBUG] Running warmup of function 6 | |
[2023-07-06 04:20:55,355] torch._dynamo.convert_frame.__recompiles: [DEBUG] ('Recompiling function forward in /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py', 'set env var TORCHDYNAMO_REPORT_GUARD_FAILURES=1 to debug further') | |
[2023-07-06 04:20:55,356] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo start tracing forward /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:190 | |
[2023-07-06 04:20:55,357] torch._dynamo.variables.builder: [DEBUG] automatic dynamic L['hidden_states'] size(1) 1 != 1000 | |
[2023-07-06 04:20:55,357] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['hidden_states'] (1, 1, 16) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None] | |
[2023-07-06 04:20:55,358] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['attention_mask'] (1, 1003) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None] | |
[2023-07-06 04:20:55,358] torch._dynamo.variables.builder: [DEBUG] automatic dynamic L['position_ids'] size(1) 1 != 1000 | |
[2023-07-06 04:20:55,358] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['position_ids'] (1, 1) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None] | |
[2023-07-06 04:20:55,359] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['past_key_value'] (2, 1, 4, 1003, 4) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None, None, None] | |
[2023-07-06 04:20:55,360] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['valid_past_index'] () [] [] | |
[2023-07-06 04:20:55,361] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:199 | |
[2023-07-06 04:20:55,361] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST output_attentions [] | |
[2023-07-06 04:20:55,361] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST True [ConstantVariable(bool)] | |
[2023-07-06 04:20:55,361] torch._dynamo.symbolic_convert: [DEBUG] TRACE IS_OP 0 [ConstantVariable(bool), ConstantVariable(bool)] | |
[2023-07-06 04:20:55,361] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(bool)] | |
[2023-07-06 04:20:55,361] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:202 | |
[2023-07-06 04:20:55,361] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:55,361] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR size [TensorVariable()] | |
[2023-07-06 04:20:55,361] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [GetAttrVariable(TensorVariable(), size)] | |
[2023-07-06 04:20:55,362] torch._dynamo.symbolic_convert: [DEBUG] TRACE UNPACK_SEQUENCE 3 [SizeVariable()] | |
[2023-07-06 04:20:55,362] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST bsz [ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:55,362] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST q_len [ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:55,362] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST _ [ConstantVariable(int)] | |
[2023-07-06 04:20:55,362] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:205 | |
[2023-07-06 04:20:55,362] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:55,362] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR qkv_proj [NNModuleVariable()] | |
[2023-07-06 04:20:55,362] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [NNModuleVariable()] | |
[2023-07-06 04:20:55,362] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,365] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR view [TensorVariable()] | |
[2023-07-06 04:20:55,365] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST bsz [GetAttrVariable(TensorVariable(), view)] | |
[2023-07-06 04:20:55,365] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST q_len [GetAttrVariable(TensorVariable(), view), ConstantVariable(int)] | |
[2023-07-06 04:20:55,365] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 3 [GetAttrVariable(TensorVariable(), view), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:55,365] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [GetAttrVariable(TensorVariable(), view), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:55,365] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR num_heads [GetAttrVariable(TensorVariable(), view), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), NNModuleVariable()] | |
[2023-07-06 04:20:55,366] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [GetAttrVariable(TensorVariable(), view), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:55,366] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR head_dim [GetAttrVariable(TensorVariable(), view), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), NNModuleVariable()] | |
[2023-07-06 04:20:55,366] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 5 [GetAttrVariable(TensorVariable(), view), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:55,367] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR permute [TensorVariable()] | |
[2023-07-06 04:20:55,367] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [GetAttrVariable(TensorVariable(), permute)] | |
[2023-07-06 04:20:55,367] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [GetAttrVariable(TensorVariable(), permute), ConstantVariable(int)] | |
[2023-07-06 04:20:55,367] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 3 [GetAttrVariable(TensorVariable(), permute), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:55,367] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [GetAttrVariable(TensorVariable(), permute), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:55,367] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 4 [GetAttrVariable(TensorVariable(), permute), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:55,367] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 5 [GetAttrVariable(TensorVariable(), permute), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:55,368] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST query_key_value_states [TensorVariable()] | |
[2023-07-06 04:20:55,368] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:207 | |
[2023-07-06 04:20:55,368] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST query_key_value_states [] | |
[2023-07-06 04:20:55,368] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [TensorVariable()] | |
[2023-07-06 04:20:55,368] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:55,369] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST query_states [TensorVariable()] | |
[2023-07-06 04:20:55,369] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:208 | |
[2023-07-06 04:20:55,369] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST query_key_value_states [] | |
[2023-07-06 04:20:55,369] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [TensorVariable()] | |
[2023-07-06 04:20:55,369] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:55,369] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_SLICE 2 [TensorVariable(), ConstantVariable(int), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:55,370] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), SliceVariable()] | |
[2023-07-06 04:20:55,370] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST key_value_states [TensorVariable()] | |
[2023-07-06 04:20:55,370] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:211 | |
[2023-07-06 04:20:55,371] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:55,371] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR rotary_emb [NNModuleVariable()] | |
[2023-07-06 04:20:55,371] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_value_states [NNModuleVariable()] | |
[2023-07-06 04:20:55,371] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,371] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST q_len [NNModuleVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,371] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [NNModuleVariable(), TensorVariable(), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:55,372] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('seq_len',) [NNModuleVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,372] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 2 [NNModuleVariable(), TensorVariable(), TensorVariable(), ConstantVariable(tuple)] | |
[2023-07-06 04:20:55,373] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:55,374] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:55,374] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:55,375] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:55,375] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:55,375] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:55,375] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:55,375] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:55,375] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:55,376] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function LlamaRotaryEmbedding.forward at 0x7fb8d0c46290>, NNModuleVariable())] | |
[2023-07-06 04:20:55,376] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:55,376] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:55,376] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:55,376] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:55,376] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:55,376] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:55,376] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:55,376] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:55,376] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:55,377] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:55,377] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:55,377] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:55,377] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:55,377] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:55,377] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:55,377] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:55,377] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:55,377] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:55,377] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:55,378] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:55,378] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:55,378] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:55,378] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:55,378] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:55,378] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:55,378] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:55,378] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:55,378] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:55,378] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:55,378] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:55,378] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:55,378] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function LlamaRotaryEmbedding.forward at 0x7fb8d0c46290>, NNModuleVariable())] | |
[2023-07-06 04:20:55,378] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function LlamaRotaryEmbedding.forward at 0x7fb8d0c46290>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:55,378] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function LlamaRotaryEmbedding.forward at 0x7fb8d0c46290>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:55,378] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function LlamaRotaryEmbedding.forward at 0x7fb8d0c46290>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:55,379] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function LlamaRotaryEmbedding.forward at 0x7fb8d0c46290>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:55,379] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object forward at 0x7fb8cddf80e0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 107> | |
[2023-07-06 04:20:55,379] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:122 | |
[2023-07-06 04:20:55,379] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:55,379] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR cos_cached [NNModuleVariable()] | |
[2023-07-06 04:20:55,379] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['self'].rotary_emb.cos_cached (1, 1, 2048, 4) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None, None] | |
[2023-07-06 04:20:55,380] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable()] | |
[2023-07-06 04:20:55,380] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:55,380] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR dtype [GetAttrVariable(TensorVariable(), to), TensorVariable()] | |
[2023-07-06 04:20:55,381] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('dtype',) [GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16)] | |
[2023-07-06 04:20:55,381] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 1 [GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16), ConstantVariable(tuple)] | |
[2023-07-06 04:20:55,382] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:123 | |
[2023-07-06 04:20:55,382] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [TensorVariable()] | |
[2023-07-06 04:20:55,382] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR sin_cached [TensorVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:55,382] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['self'].rotary_emb.sin_cached (1, 1, 2048, 4) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None, None] | |
[2023-07-06 04:20:55,383] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,383] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TensorVariable(), GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:55,383] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR dtype [TensorVariable(), GetAttrVariable(TensorVariable(), to), TensorVariable()] | |
[2023-07-06 04:20:55,383] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('dtype',) [TensorVariable(), GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16)] | |
[2023-07-06 04:20:55,383] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 1 [TensorVariable(), GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16), ConstantVariable(tuple)] | |
[2023-07-06 04:20:55,385] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:121 | |
[2023-07-06 04:20:55,385] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_TUPLE 2 [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,385] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TupleVariable()] | |
[2023-07-06 04:20:55,385] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object forward at 0x7fb8cddf80e0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 107> | |
[2023-07-06 04:20:55,385] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TupleVariable()] | |
[2023-07-06 04:20:55,385] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:55,385] torch._dynamo.symbolic_convert: [DEBUG] TRACE UNPACK_SEQUENCE 2 [TupleVariable()] | |
[2023-07-06 04:20:55,385] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST cos [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,385] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST sin [TensorVariable()] | |
[2023-07-06 04:20:55,385] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:212 | |
[2023-07-06 04:20:55,385] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL apply_rotary_pos_emb_opt [] | |
[2023-07-06 04:20:55,385] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST query_states [UserFunctionVariable()] | |
[2023-07-06 04:20:55,385] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_value_states [UserFunctionVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,385] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [UserFunctionVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,385] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [UserFunctionVariable(), TensorVariable(), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:55,386] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST cos [UserFunctionVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,386] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST sin [UserFunctionVariable(), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,386] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST position_ids [UserFunctionVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,386] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 5 [UserFunctionVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,386] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object apply_rotary_pos_emb_opt at 0x7fb8cddf82f0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 135> | |
[2023-07-06 04:20:55,387] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:139 | |
[2023-07-06 04:20:55,387] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST cos [] | |
[2023-07-06 04:20:55,387] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR squeeze [TensorVariable()] | |
[2023-07-06 04:20:55,387] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [GetAttrVariable(TensorVariable(), squeeze)] | |
[2023-07-06 04:20:55,387] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), squeeze), ConstantVariable(int)] | |
[2023-07-06 04:20:55,388] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR squeeze [TensorVariable()] | |
[2023-07-06 04:20:55,388] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [GetAttrVariable(TensorVariable(), squeeze)] | |
[2023-07-06 04:20:55,388] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), squeeze), ConstantVariable(int)] | |
[2023-07-06 04:20:55,389] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST cos [TensorVariable()] | |
[2023-07-06 04:20:55,389] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:140 | |
[2023-07-06 04:20:55,389] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST sin [] | |
[2023-07-06 04:20:55,389] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR squeeze [TensorVariable()] | |
[2023-07-06 04:20:55,389] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [GetAttrVariable(TensorVariable(), squeeze)] | |
[2023-07-06 04:20:55,389] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), squeeze), ConstantVariable(int)] | |
[2023-07-06 04:20:55,390] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR squeeze [TensorVariable()] | |
[2023-07-06 04:20:55,390] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [GetAttrVariable(TensorVariable(), squeeze)] | |
[2023-07-06 04:20:55,390] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), squeeze), ConstantVariable(int)] | |
[2023-07-06 04:20:55,391] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST sin [TensorVariable()] | |
[2023-07-06 04:20:55,391] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:141 | |
[2023-07-06 04:20:55,391] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST cos [] | |
[2023-07-06 04:20:55,391] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST position_ids [TensorVariable()] | |
[2023-07-06 04:20:55,391] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,392] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR unsqueeze [TensorVariable()] | |
[2023-07-06 04:20:55,393] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [GetAttrVariable(TensorVariable(), unsqueeze)] | |
[2023-07-06 04:20:55,393] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), unsqueeze), ConstantVariable(int)] | |
[2023-07-06 04:20:55,393] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST cos [TensorVariable()] | |
[2023-07-06 04:20:55,393] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:142 | |
[2023-07-06 04:20:55,393] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST sin [] | |
[2023-07-06 04:20:55,393] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST position_ids [TensorVariable()] | |
[2023-07-06 04:20:55,393] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,395] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR unsqueeze [TensorVariable()] | |
[2023-07-06 04:20:55,395] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [GetAttrVariable(TensorVariable(), unsqueeze)] | |
[2023-07-06 04:20:55,395] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), unsqueeze), ConstantVariable(int)] | |
[2023-07-06 04:20:55,396] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST sin [TensorVariable()] | |
[2023-07-06 04:20:55,396] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:143 | |
[2023-07-06 04:20:55,396] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST q [] | |
[2023-07-06 04:20:55,396] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST cos [TensorVariable()] | |
[2023-07-06 04:20:55,396] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,397] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL rotate_half [TensorVariable()] | |
[2023-07-06 04:20:55,397] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST q [TensorVariable(), UserFunctionVariable()] | |
[2023-07-06 04:20:55,397] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [TensorVariable(), UserFunctionVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,397] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object rotate_half at 0x7fb8cddf8240, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 127> | |
[2023-07-06 04:20:55,397] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:129 | |
[2023-07-06 04:20:55,397] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:55,398] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR roll [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:55,398] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TorchVariable(<built-in method roll of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:55,398] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:55,398] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR shape [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,398] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ShapeVariable()] | |
[2023-07-06 04:20:55,399] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ShapeVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:55,399] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:55,399] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_FLOOR_DIVIDE None [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:55,399] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:55,399] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('shifts', 'dims') [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:55,399] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 3 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int), ConstantVariable(int), ConstantVariable(tuple)] | |
[2023-07-06 04:20:55,400] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST x [TensorVariable()] | |
[2023-07-06 04:20:55,400] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:130 | |
[2023-07-06 04:20:55,400] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [] | |
[2023-07-06 04:20:55,400] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST Ellipsis [TensorVariable()] | |
[2023-07-06 04:20:55,400] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [TensorVariable(), ConstantVariable(ellipsis)] | |
[2023-07-06 04:20:55,400] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:55,400] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR shape [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), TensorVariable()] | |
[2023-07-06 04:20:55,401] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ShapeVariable()] | |
[2023-07-06 04:20:55,401] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ShapeVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:55,401] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ConstantVariable(int)] | |
[2023-07-06 04:20:55,401] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_FLOOR_DIVIDE None [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:55,401] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_SLICE 2 [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ConstantVariable(int)] | |
[2023-07-06 04:20:55,401] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_TUPLE 2 [TensorVariable(), ConstantVariable(ellipsis), SliceVariable()] | |
[2023-07-06 04:20:55,401] torch._dynamo.symbolic_convert: [DEBUG] TRACE DUP_TOP_TWO None [TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:55,401] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), TupleVariable(), TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:55,402] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TensorVariable(), TupleVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,402] torch._dynamo.symbolic_convert: [DEBUG] TRACE INPLACE_MULTIPLY None [TensorVariable(), TupleVariable(), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:55,403] torch._dynamo.symbolic_convert: [DEBUG] TRACE ROT_THREE None [TensorVariable(), TupleVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,403] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_SUBSCR None [TensorVariable(), TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:55,403] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:131 | |
[2023-07-06 04:20:55,403] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [] | |
[2023-07-06 04:20:55,403] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:55,403] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object rotate_half at 0x7fb8cddf8240, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 127> | |
[2023-07-06 04:20:55,403] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST sin [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,403] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,404] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,405] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST q_embed [TensorVariable()] | |
[2023-07-06 04:20:55,405] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:144 | |
[2023-07-06 04:20:55,405] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_states [] | |
[2023-07-06 04:20:55,405] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR copy_ [TensorVariable()] | |
[2023-07-06 04:20:55,406] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_states [GetAttrVariable(TensorVariable(), copy_)] | |
[2023-07-06 04:20:55,406] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST cos [GetAttrVariable(TensorVariable(), copy_), TensorVariable()] | |
[2023-07-06 04:20:55,406] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [GetAttrVariable(TensorVariable(), copy_), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,407] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL rotate_half [GetAttrVariable(TensorVariable(), copy_), TensorVariable()] | |
[2023-07-06 04:20:55,407] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_states [GetAttrVariable(TensorVariable(), copy_), TensorVariable(), UserFunctionVariable()] | |
[2023-07-06 04:20:55,407] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), copy_), TensorVariable(), UserFunctionVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,407] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object rotate_half at 0x7fb8cddf8240, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 127> | |
[2023-07-06 04:20:55,407] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:129 | |
[2023-07-06 04:20:55,407] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:55,407] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR roll [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:55,408] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TorchVariable(<built-in method roll of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:55,408] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:55,408] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR shape [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,408] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ShapeVariable()] | |
[2023-07-06 04:20:55,408] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ShapeVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:55,408] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:55,408] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_FLOOR_DIVIDE None [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:55,408] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:55,408] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('shifts', 'dims') [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:55,409] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 3 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int), ConstantVariable(int), ConstantVariable(tuple)] | |
[2023-07-06 04:20:55,409] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST x [TensorVariable()] | |
[2023-07-06 04:20:55,409] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:130 | |
[2023-07-06 04:20:55,409] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [] | |
[2023-07-06 04:20:55,409] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST Ellipsis [TensorVariable()] | |
[2023-07-06 04:20:55,410] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [TensorVariable(), ConstantVariable(ellipsis)] | |
[2023-07-06 04:20:55,410] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:55,410] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR shape [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), TensorVariable()] | |
[2023-07-06 04:20:55,410] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ShapeVariable()] | |
[2023-07-06 04:20:55,410] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ShapeVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:55,410] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ConstantVariable(int)] | |
[2023-07-06 04:20:55,410] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_FLOOR_DIVIDE None [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:55,410] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_SLICE 2 [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ConstantVariable(int)] | |
[2023-07-06 04:20:55,411] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_TUPLE 2 [TensorVariable(), ConstantVariable(ellipsis), SliceVariable()] | |
[2023-07-06 04:20:55,411] torch._dynamo.symbolic_convert: [DEBUG] TRACE DUP_TOP_TWO None [TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:55,411] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), TupleVariable(), TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:55,412] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TensorVariable(), TupleVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,412] torch._dynamo.symbolic_convert: [DEBUG] TRACE INPLACE_MULTIPLY None [TensorVariable(), TupleVariable(), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:55,412] torch._dynamo.symbolic_convert: [DEBUG] TRACE ROT_THREE None [TensorVariable(), TupleVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,412] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_SUBSCR None [TensorVariable(), TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:55,413] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:131 | |
[2023-07-06 04:20:55,413] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [] | |
[2023-07-06 04:20:55,413] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:55,413] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object rotate_half at 0x7fb8cddf8240, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 127> | |
[2023-07-06 04:20:55,413] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST sin [GetAttrVariable(TensorVariable(), copy_), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,413] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [GetAttrVariable(TensorVariable(), copy_), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,414] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [GetAttrVariable(TensorVariable(), copy_), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,415] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), copy_), TensorVariable()] | |
[2023-07-06 04:20:55,415] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_TOP None [TensorVariable()] | |
[2023-07-06 04:20:55,415] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:145 | |
[2023-07-06 04:20:55,415] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST q_embed [] | |
[2023-07-06 04:20:55,415] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:55,415] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object apply_rotary_pos_emb_opt at 0x7fb8cddf82f0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 135> | |
[2023-07-06 04:20:55,415] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST query_states [TensorVariable()] | |
[2023-07-06 04:20:55,415] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:215 | |
[2023-07-06 04:20:55,415] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [] | |
[2023-07-06 04:20:55,415] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST past_kv_slice_start [TensorVariable()] | |
[2023-07-06 04:20:55,416] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:216 | |
[2023-07-06 04:20:55,416] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:55,416] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR eq [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:55,416] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [TorchVariable(<built-in method eq of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:55,416] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [TorchVariable(<built-in method eq of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:55,416] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 2 [TorchVariable(<built-in method eq of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:55,417] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR int [TensorVariable()] | |
[2023-07-06 04:20:55,417] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [GetAttrVariable(TensorVariable(), int)] | |
[2023-07-06 04:20:55,418] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST q_len [TensorVariable()] | |
[2023-07-06 04:20:55,418] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:55,419] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [TensorVariable()] | |
[2023-07-06 04:20:55,419] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR ne [TensorVariable(), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:55,419] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [TensorVariable(), TorchVariable(<built-in method ne of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:55,419] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [TensorVariable(), TorchVariable(<built-in method ne of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:55,419] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 2 [TensorVariable(), TorchVariable(<built-in method ne of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:55,420] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR int [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,420] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TensorVariable(), GetAttrVariable(TensorVariable(), int)] | |
[2023-07-06 04:20:55,421] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,421] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,422] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TensorVariable(), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:55,422] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,423] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,424] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST past_kv_slice_end [TensorVariable()] | |
[2023-07-06 04:20:55,424] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:217 | |
[2023-07-06 04:20:55,424] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:55,424] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR eq [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:55,424] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [TorchVariable(<built-in method eq of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:55,424] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [TorchVariable(<built-in method eq of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:55,424] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 2 [TorchVariable(<built-in method eq of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:55,425] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR int [TensorVariable()] | |
[2023-07-06 04:20:55,425] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [GetAttrVariable(TensorVariable(), int)] | |
[2023-07-06 04:20:55,426] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_value_states [TensorVariable()] | |
[2023-07-06 04:20:55,426] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR shape [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,427] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -2 [TensorVariable(), ShapeVariable()] | |
[2023-07-06 04:20:55,427] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), ShapeVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:55,427] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:55,428] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [TensorVariable()] | |
[2023-07-06 04:20:55,428] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR ne [TensorVariable(), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:55,428] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [TensorVariable(), TorchVariable(<built-in method ne of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:55,428] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [TensorVariable(), TorchVariable(<built-in method ne of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:55,428] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 2 [TensorVariable(), TorchVariable(<built-in method ne of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:55,429] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR int [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,429] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TensorVariable(), GetAttrVariable(TensorVariable(), int)] | |
[2023-07-06 04:20:55,430] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST past_kv_slice_end [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,430] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,431] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,432] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST past_state_slice_end [TensorVariable()] | |
[2023-07-06 04:20:55,432] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:218 | |
[2023-07-06 04:20:55,432] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_value_states [] | |
[2023-07-06 04:20:55,432] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST past_key_value [TensorVariable()] | |
[2023-07-06 04:20:55,432] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST Ellipsis [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,432] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST past_kv_slice_start [TensorVariable(), TensorVariable(), ConstantVariable(ellipsis)] | |
[2023-07-06 04:20:55,432] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST past_kv_slice_end [TensorVariable(), TensorVariable(), ConstantVariable(ellipsis), TensorVariable()] | |
[2023-07-06 04:20:55,432] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_SLICE 2 [TensorVariable(), TensorVariable(), ConstantVariable(ellipsis), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:55,432] torch._dynamo.symbolic_convert: [DEBUG] step triggered compile | |
Traceback (most recent call last): | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 668, in step | |
getattr(self, inst.opname)(inst) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 1251, in BUILD_SLICE | |
SliceVariable( | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/base.py", line 27, in __call__ | |
obj = type.__call__(cls, *args, **kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/lists.py", line 616, in __init__ | |
unimplemented("Dynamic slicing on data-dependent value is not supported") | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/exc.py", line 140, in unimplemented | |
raise Unsupported(msg) | |
torch._dynamo.exc.Unsupported: Dynamic slicing on data-dependent value is not supported | |
[2023-07-06 04:20:55,432] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:55,432] torch._dynamo.output_graph: [DEBUG] COMPILING GRAPH due to GraphCompileReason(reason='step_unsupported', user_stack=[<FrameSummary file /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py, line 218 in forward>], graph_break=True) | |
[2023-07-06 04:20:55,434] torch._dynamo.output_graph: [DEBUG] REMOVE UNUSED GRAPHARG L['attention_mask'] | |
[2023-07-06 04:20:55,434] torch._dynamo.output_graph: [DEBUG] REMOVE UNUSED GRAPHARG L['past_key_value'] | |
[2023-07-06 04:20:55,436] torch._dynamo.output_graph.__graph_code: [DEBUG] TRACED GRAPH | |
===== __compiled_fn_17 ===== | |
<eval_with_key>.182 class GraphModule(torch.nn.Module): | |
def forward(self, L_hidden_states_ : torch.Tensor, L_position_ids_ : torch.Tensor, L_valid_past_index_ : torch.Tensor): | |
l_hidden_states_ = L_hidden_states_ | |
l_position_ids_ = L_position_ids_ | |
l_valid_past_index_ = L_valid_past_index_ | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:205, code: query_key_value_states = self.qkv_proj(hidden_states).view(bsz, q_len, 3, self.num_heads, self.head_dim).permute(2, 0, 3, 1, 4) | |
l__self___qkv_proj = self.L__self___qkv_proj(l_hidden_states_); l_hidden_states_ = None | |
view = l__self___qkv_proj.view(1, 1, 3, 4, 4); l__self___qkv_proj = None | |
permute = view.permute(2, 0, 3, 1, 4); view = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:207, code: query_states = query_key_value_states[0] | |
getitem = permute[0] | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:208, code: key_value_states = query_key_value_states[1:] | |
getitem_1 = permute[slice(1, None, None)]; permute = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:211, code: cos, sin = self.rotary_emb(key_value_states, seq_len=valid_past_index + q_len) | |
add = l_valid_past_index_ + 1 | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:122, code: self.cos_cached.to(dtype=x.dtype), | |
l__self___rotary_emb_cos_cached = self.L__self___rotary_emb_cos_cached | |
to = l__self___rotary_emb_cos_cached.to(dtype = torch.float16); l__self___rotary_emb_cos_cached = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:123, code: self.sin_cached.to(dtype=x.dtype), | |
l__self___rotary_emb_sin_cached = self.L__self___rotary_emb_sin_cached | |
to_1 = l__self___rotary_emb_sin_cached.to(dtype = torch.float16); l__self___rotary_emb_sin_cached = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:212, code: query_states = apply_rotary_pos_emb_opt(query_states, key_value_states[0], cos, sin, position_ids) | |
getitem_2 = getitem_1[0] | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:139, code: cos = cos.squeeze(1).squeeze(0) # [seq_len, dim] | |
squeeze = to.squeeze(1); to = None | |
squeeze_1 = squeeze.squeeze(0); squeeze = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:140, code: sin = sin.squeeze(1).squeeze(0) # [seq_len, dim] | |
squeeze_2 = to_1.squeeze(1); to_1 = None | |
squeeze_3 = squeeze_2.squeeze(0); squeeze_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:141, code: cos = cos[position_ids].unsqueeze(1) # [bs, 1, seq_len, dim] | |
getitem_3 = squeeze_1[l_position_ids_]; squeeze_1 = None | |
unsqueeze = getitem_3.unsqueeze(1); getitem_3 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:142, code: sin = sin[position_ids].unsqueeze(1) # [bs, 1, seq_len, dim] | |
getitem_4 = squeeze_3[l_position_ids_]; squeeze_3 = l_position_ids_ = None | |
unsqueeze_1 = getitem_4.unsqueeze(1); getitem_4 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:143, code: q_embed = (q * cos) + (rotate_half(q) * sin) | |
mul = getitem * unsqueeze | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:129, code: x = torch.roll(x, shifts=x.shape[-1] // 2, dims=-1) | |
roll = torch.roll(getitem, shifts = 2, dims = -1); getitem = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:130, code: x[..., : x.shape[-1] // 2] *= -1 | |
getitem_5 = roll[(Ellipsis, slice(None, 2, None))] | |
getitem_5 *= -1; imul = getitem_5; getitem_5 = None | |
roll[(Ellipsis, slice(None, 2, None))] = imul; setitem = roll; imul = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:143, code: q_embed = (q * cos) + (rotate_half(q) * sin) | |
mul_1 = roll * unsqueeze_1; roll = None | |
add_1 = mul + mul_1; mul = mul_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:144, code: key_states.copy_((key_states * cos) + (rotate_half(key_states) * sin)) | |
mul_2 = getitem_2 * unsqueeze; unsqueeze = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:129, code: x = torch.roll(x, shifts=x.shape[-1] // 2, dims=-1) | |
roll_1 = torch.roll(getitem_2, shifts = 2, dims = -1) | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:130, code: x[..., : x.shape[-1] // 2] *= -1 | |
getitem_6 = roll_1[(Ellipsis, slice(None, 2, None))] | |
getitem_6 *= -1; imul_1 = getitem_6; getitem_6 = None | |
roll_1[(Ellipsis, slice(None, 2, None))] = imul_1; setitem_1 = roll_1; imul_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:144, code: key_states.copy_((key_states * cos) + (rotate_half(key_states) * sin)) | |
mul_3 = roll_1 * unsqueeze_1; roll_1 = unsqueeze_1 = None | |
add_2 = mul_2 + mul_3; mul_2 = mul_3 = None | |
copy_ = getitem_2.copy_(add_2); getitem_2 = add_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:216, code: past_kv_slice_end = torch.eq(valid_past_index, 0).int() * q_len + torch.ne(valid_past_index, 0).int() * (valid_past_index + 1) | |
eq = torch.eq(l_valid_past_index_, 0) | |
int_1 = eq.int(); eq = None | |
mul_4 = int_1 * 1; int_1 = None | |
ne = torch.ne(l_valid_past_index_, 0) | |
int_2 = ne.int(); ne = None | |
add_3 = l_valid_past_index_ + 1 | |
mul_5 = int_2 * add_3; int_2 = add_3 = None | |
add_4 = mul_4 + mul_5; mul_4 = mul_5 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:217, code: past_state_slice_end = torch.eq(valid_past_index, 0).int() * key_value_states.shape[-2] + torch.ne(valid_past_index, 0).int() * (past_kv_slice_end) | |
eq_1 = torch.eq(l_valid_past_index_, 0) | |
int_3 = eq_1.int(); eq_1 = None | |
mul_6 = int_3 * 1; int_3 = None | |
ne_1 = torch.ne(l_valid_past_index_, 0); l_valid_past_index_ = None | |
int_4 = ne_1.int(); ne_1 = None | |
mul_7 = int_4 * add_4; int_4 = None | |
add_5 = mul_6 + mul_7; mul_6 = mul_7 = None | |
return (add_1, getitem_1, add_4, add_5) | |
[2023-07-06 04:20:55,437] torch._dynamo.output_graph.__graph: [DEBUG] TRACED GRAPH | |
__compiled_fn_17 <eval_with_key>.182 opcode name target args kwargs | |
------------- ------------------------------- ------------------------------------------------------- -------------------------------------------------- ------------------------- | |
placeholder l_hidden_states_ L_hidden_states_ () {} | |
placeholder l_position_ids_ L_position_ids_ () {} | |
placeholder l_valid_past_index_ L_valid_past_index_ () {} | |
call_module l__self___qkv_proj L__self___qkv_proj (l_hidden_states_,) {} | |
call_method view view (l__self___qkv_proj, 1, 1, 3, 4, 4) {} | |
call_method permute permute (view, 2, 0, 3, 1, 4) {} | |
call_function getitem <built-in function getitem> (permute, 0) {} | |
call_function getitem_1 <built-in function getitem> (permute, slice(1, None, None)) {} | |
call_function add <built-in function add> (l_valid_past_index_, 1) {} | |
get_attr l__self___rotary_emb_cos_cached L__self___rotary_emb_cos_cached () {} | |
call_method to to (l__self___rotary_emb_cos_cached,) {'dtype': torch.float16} | |
get_attr l__self___rotary_emb_sin_cached L__self___rotary_emb_sin_cached () {} | |
call_method to_1 to (l__self___rotary_emb_sin_cached,) {'dtype': torch.float16} | |
call_function getitem_2 <built-in function getitem> (getitem_1, 0) {} | |
call_method squeeze squeeze (to, 1) {} | |
call_method squeeze_1 squeeze (squeeze, 0) {} | |
call_method squeeze_2 squeeze (to_1, 1) {} | |
call_method squeeze_3 squeeze (squeeze_2, 0) {} | |
call_function getitem_3 <built-in function getitem> (squeeze_1, l_position_ids_) {} | |
call_method unsqueeze unsqueeze (getitem_3, 1) {} | |
call_function getitem_4 <built-in function getitem> (squeeze_3, l_position_ids_) {} | |
call_method unsqueeze_1 unsqueeze (getitem_4, 1) {} | |
call_function mul <built-in function mul> (getitem, unsqueeze) {} | |
call_function roll <built-in method roll of type object at 0x7fb940469580> (getitem,) {'shifts': 2, 'dims': -1} | |
call_function getitem_5 <built-in function getitem> (roll, (Ellipsis, slice(None, 2, None))) {} | |
call_function imul <built-in function imul> (getitem_5, -1) {} | |
call_function setitem <built-in function setitem> (roll, (Ellipsis, slice(None, 2, None)), imul) {} | |
call_function mul_1 <built-in function mul> (roll, unsqueeze_1) {} | |
call_function add_1 <built-in function add> (mul, mul_1) {} | |
call_function mul_2 <built-in function mul> (getitem_2, unsqueeze) {} | |
call_function roll_1 <built-in method roll of type object at 0x7fb940469580> (getitem_2,) {'shifts': 2, 'dims': -1} | |
call_function getitem_6 <built-in function getitem> (roll_1, (Ellipsis, slice(None, 2, None))) {} | |
call_function imul_1 <built-in function imul> (getitem_6, -1) {} | |
call_function setitem_1 <built-in function setitem> (roll_1, (Ellipsis, slice(None, 2, None)), imul_1) {} | |
call_function mul_3 <built-in function mul> (roll_1, unsqueeze_1) {} | |
call_function add_2 <built-in function add> (mul_2, mul_3) {} | |
call_method copy_ copy_ (getitem_2, add_2) {} | |
call_function eq <built-in method eq of type object at 0x7fb940469580> (l_valid_past_index_, 0) {} | |
call_method int_1 int (eq,) {} | |
call_function mul_4 <built-in function mul> (int_1, 1) {} | |
call_function ne <built-in method ne of type object at 0x7fb940469580> (l_valid_past_index_, 0) {} | |
call_method int_2 int (ne,) {} | |
call_function add_3 <built-in function add> (l_valid_past_index_, 1) {} | |
call_function mul_5 <built-in function mul> (int_2, add_3) {} | |
call_function add_4 <built-in function add> (mul_4, mul_5) {} | |
call_function eq_1 <built-in method eq of type object at 0x7fb940469580> (l_valid_past_index_, 0) {} | |
call_method int_3 int (eq_1,) {} | |
call_function mul_6 <built-in function mul> (int_3, 1) {} | |
call_function ne_1 <built-in method ne of type object at 0x7fb940469580> (l_valid_past_index_, 0) {} | |
call_method int_4 int (ne_1,) {} | |
call_function mul_7 <built-in function mul> (int_4, add_4) {} | |
call_function add_5 <built-in function add> (mul_6, mul_7) {} | |
output output output ((add_1, getitem_1, add_4, add_5),) {} | |
[2023-07-06 04:20:55,439] torch._dynamo.output_graph: [INFO] Step 2: calling compiler function inductor | |
[2023-07-06 04:20:55,561] torch._functorch.aot_autograd.__aot_graphs: [INFO] TRACED GRAPH | |
===== Forward graph 22 ===== | |
<eval_with_key>.186 class <lambda>(torch.nn.Module): | |
def forward(self, arg0_1: f16[48, 16], arg1_1: f32[1, 1, 2048, 4], arg2_1: f32[1, 1, 2048, 4], arg3_1: f16[1, 1, 16], arg4_1: i64[1, 1], arg5_1: i64[]): | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:205, code: query_key_value_states = self.qkv_proj(hidden_states).view(bsz, q_len, 3, self.num_heads, self.head_dim).permute(2, 0, 3, 1, 4) | |
permute: f16[16, 48] = torch.ops.aten.permute.default(arg0_1, [1, 0]); arg0_1 = None | |
view: f16[1, 16] = torch.ops.aten.view.default(arg3_1, [1, 16]); arg3_1 = None | |
mm: f16[1, 48] = torch.ops.aten.mm.default(view, permute); view = permute = None | |
view_1: f16[1, 1, 48] = torch.ops.aten.view.default(mm, [1, 1, 48]) | |
view_2: f16[1, 1, 3, 4, 4] = torch.ops.aten.view.default(view_1, [1, 1, 3, 4, 4]); view_1 = None | |
permute_1: f16[3, 1, 4, 1, 4] = torch.ops.aten.permute.default(view_2, [2, 0, 3, 1, 4]); view_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:207, code: query_states = query_key_value_states[0] | |
select: f16[1, 4, 1, 4] = torch.ops.aten.select.int(permute_1, 0, 0) | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:208, code: key_value_states = query_key_value_states[1:] | |
slice_1: f16[2, 1, 4, 1, 4] = torch.ops.aten.slice.Tensor(permute_1, 0, 1, 9223372036854775807); permute_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:122, code: self.cos_cached.to(dtype=x.dtype), | |
convert_element_type: f16[1, 1, 2048, 4] = torch.ops.prims.convert_element_type.default(arg1_1, torch.float16); arg1_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:123, code: self.sin_cached.to(dtype=x.dtype), | |
convert_element_type_1: f16[1, 1, 2048, 4] = torch.ops.prims.convert_element_type.default(arg2_1, torch.float16); arg2_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:212, code: query_states = apply_rotary_pos_emb_opt(query_states, key_value_states[0], cos, sin, position_ids) | |
select_1: f16[1, 4, 1, 4] = torch.ops.aten.select.int(slice_1, 0, 0); slice_1 = None | |
# No stacktrace found for following nodes | |
squeeze: f16[1, 2048, 4] = torch.ops.aten.squeeze.dim(convert_element_type, 1); convert_element_type = None | |
squeeze_1: f16[2048, 4] = torch.ops.aten.squeeze.dim(squeeze, 0); squeeze = None | |
squeeze_2: f16[1, 2048, 4] = torch.ops.aten.squeeze.dim(convert_element_type_1, 1); convert_element_type_1 = None | |
squeeze_3: f16[2048, 4] = torch.ops.aten.squeeze.dim(squeeze_2, 0); squeeze_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:141, code: cos = cos[position_ids].unsqueeze(1) # [bs, 1, seq_len, dim] | |
index: f16[1, 1, 4] = torch.ops.aten.index.Tensor(squeeze_1, [arg4_1]); squeeze_1 = None | |
unsqueeze: f16[1, 1, 1, 4] = torch.ops.aten.unsqueeze.default(index, 1); index = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:142, code: sin = sin[position_ids].unsqueeze(1) # [bs, 1, seq_len, dim] | |
index_1: f16[1, 1, 4] = torch.ops.aten.index.Tensor(squeeze_3, [arg4_1]); squeeze_3 = arg4_1 = None | |
unsqueeze_1: f16[1, 1, 1, 4] = torch.ops.aten.unsqueeze.default(index_1, 1); index_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:143, code: q_embed = (q * cos) + (rotate_half(q) * sin) | |
mul: f16[1, 4, 1, 4] = torch.ops.aten.mul.Tensor(select, unsqueeze) | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:129, code: x = torch.roll(x, shifts=x.shape[-1] // 2, dims=-1) | |
roll: f16[1, 4, 1, 4] = torch.ops.aten.roll.default(select, [2], [-1]); select = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:130, code: x[..., : x.shape[-1] // 2] *= -1 | |
slice_2: f16[1, 4, 1, 2] = torch.ops.aten.slice.Tensor(roll, 3, 0, 2) | |
mul_1: f16[1, 4, 1, 2] = torch.ops.aten.mul.Tensor(slice_2, -1); slice_2 = None | |
slice_scatter: f16[1, 4, 1, 4] = torch.ops.aten.slice_scatter.default(roll, mul_1, 3, 0, 2); roll = mul_1 = None | |
slice_3: f16[1, 4, 1, 2] = torch.ops.aten.slice.Tensor(slice_scatter, 3, 0, 2) | |
slice_scatter_1: f16[1, 4, 1, 4] = torch.ops.aten.slice_scatter.default(slice_scatter, slice_3, 3, 0, 2); slice_scatter = slice_3 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:143, code: q_embed = (q * cos) + (rotate_half(q) * sin) | |
mul_2: f16[1, 4, 1, 4] = torch.ops.aten.mul.Tensor(slice_scatter_1, unsqueeze_1); slice_scatter_1 = None | |
add_1: f16[1, 4, 1, 4] = torch.ops.aten.add.Tensor(mul, mul_2); mul = mul_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:144, code: key_states.copy_((key_states * cos) + (rotate_half(key_states) * sin)) | |
mul_3: f16[1, 4, 1, 4] = torch.ops.aten.mul.Tensor(select_1, unsqueeze); unsqueeze = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:129, code: x = torch.roll(x, shifts=x.shape[-1] // 2, dims=-1) | |
roll_1: f16[1, 4, 1, 4] = torch.ops.aten.roll.default(select_1, [2], [-1]); select_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:130, code: x[..., : x.shape[-1] // 2] *= -1 | |
slice_7: f16[1, 4, 1, 2] = torch.ops.aten.slice.Tensor(roll_1, 3, 0, 2) | |
mul_4: f16[1, 4, 1, 2] = torch.ops.aten.mul.Tensor(slice_7, -1); slice_7 = None | |
slice_scatter_2: f16[1, 4, 1, 4] = torch.ops.aten.slice_scatter.default(roll_1, mul_4, 3, 0, 2); roll_1 = mul_4 = None | |
slice_8: f16[1, 4, 1, 2] = torch.ops.aten.slice.Tensor(slice_scatter_2, 3, 0, 2) | |
slice_scatter_3: f16[1, 4, 1, 4] = torch.ops.aten.slice_scatter.default(slice_scatter_2, slice_8, 3, 0, 2); slice_scatter_2 = slice_8 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:144, code: key_states.copy_((key_states * cos) + (rotate_half(key_states) * sin)) | |
mul_5: f16[1, 4, 1, 4] = torch.ops.aten.mul.Tensor(slice_scatter_3, unsqueeze_1); slice_scatter_3 = unsqueeze_1 = None | |
add_2: f16[1, 4, 1, 4] = torch.ops.aten.add.Tensor(mul_3, mul_5); mul_3 = mul_5 = None | |
view_3: f16[1, 1, 48] = torch.ops.aten.view.default(mm, [1, 1, 48]); mm = None | |
view_4: f16[1, 1, 3, 4, 4] = torch.ops.aten.view.default(view_3, [1, 1, 3, 4, 4]); view_3 = None | |
permute_2: f16[3, 1, 4, 1, 4] = torch.ops.aten.permute.default(view_4, [2, 0, 3, 1, 4]); view_4 = None | |
slice_12: f16[2, 1, 4, 1, 4] = torch.ops.aten.slice.Tensor(permute_2, 0, 1, 9223372036854775807) | |
select_scatter: f16[2, 1, 4, 1, 4] = torch.ops.aten.select_scatter.default(slice_12, add_2, 0, 0); slice_12 = add_2 = None | |
slice_scatter_4: f16[3, 1, 4, 1, 4] = torch.ops.aten.slice_scatter.default(permute_2, select_scatter, 0, 1, 9223372036854775807); permute_2 = select_scatter = None | |
permute_3: f16[1, 1, 3, 4, 4] = torch.ops.aten.permute.default(slice_scatter_4, [1, 3, 0, 2, 4]); slice_scatter_4 = None | |
view_5: f16[1, 1, 48] = torch.ops.aten.view.default(permute_3, [1, 1, 48]); permute_3 = None | |
view_6: f16[1, 48] = torch.ops.aten.view.default(view_5, [1, 48]); view_5 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:216, code: past_kv_slice_end = torch.eq(valid_past_index, 0).int() * q_len + torch.ne(valid_past_index, 0).int() * (valid_past_index + 1) | |
eq: b8[] = torch.ops.aten.eq.Scalar(arg5_1, 0) | |
convert_element_type_2: i32[] = torch.ops.prims.convert_element_type.default(eq, torch.int32); eq = None | |
mul_6: i32[] = torch.ops.aten.mul.Tensor(convert_element_type_2, 1); convert_element_type_2 = None | |
ne: b8[] = torch.ops.aten.ne.Scalar(arg5_1, 0) | |
convert_element_type_3: i32[] = torch.ops.prims.convert_element_type.default(ne, torch.int32); ne = None | |
add_3: i64[] = torch.ops.aten.add.Tensor(arg5_1, 1) | |
mul_7: i64[] = torch.ops.aten.mul.Tensor(convert_element_type_3, add_3); convert_element_type_3 = add_3 = None | |
add_4: i64[] = torch.ops.aten.add.Tensor(mul_6, mul_7); mul_6 = mul_7 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:217, code: past_state_slice_end = torch.eq(valid_past_index, 0).int() * key_value_states.shape[-2] + torch.ne(valid_past_index, 0).int() * (past_kv_slice_end) | |
eq_1: b8[] = torch.ops.aten.eq.Scalar(arg5_1, 0) | |
convert_element_type_4: i32[] = torch.ops.prims.convert_element_type.default(eq_1, torch.int32); eq_1 = None | |
mul_8: i32[] = torch.ops.aten.mul.Tensor(convert_element_type_4, 1); convert_element_type_4 = None | |
ne_1: b8[] = torch.ops.aten.ne.Scalar(arg5_1, 0); arg5_1 = None | |
convert_element_type_5: i32[] = torch.ops.prims.convert_element_type.default(ne_1, torch.int32); ne_1 = None | |
mul_9: i64[] = torch.ops.aten.mul.Tensor(convert_element_type_5, add_4); convert_element_type_5 = None | |
add_5: i64[] = torch.ops.aten.add.Tensor(mul_8, mul_9); mul_8 = mul_9 = None | |
# No stacktrace found for following nodes | |
view_9: f16[1, 1, 48] = torch.ops.aten.view.default(view_6, [1, 1, 48]); view_6 = None | |
view_10: f16[1, 1, 3, 4, 4] = torch.ops.aten.view.default(view_9, [1, 1, 3, 4, 4]); view_9 = None | |
permute_5: f16[3, 1, 4, 1, 4] = torch.ops.aten.permute.default(view_10, [2, 0, 3, 1, 4]); view_10 = None | |
slice_14: f16[2, 1, 4, 1, 4] = torch.ops.aten.slice.Tensor(permute_5, 0, 1, 9223372036854775807); permute_5 = None | |
return (add_1, slice_14, add_4, add_5) | |
[2023-07-06 04:20:55,596] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 9 | |
[2023-07-06 04:20:55,706] torch._inductor.graph: [DEBUG] Force channels last inputs for 0 conv for the current graph with id 9 | |
[2023-07-06 04:20:55,887] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 1 configs | |
[2023-07-06 04:20:55,887] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 32, num_warps: 1, num_stages: 1 | |
[2023-07-06 04:20:55,888] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 1 configs | |
[2023-07-06 04:20:55,888] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 16, num_warps: 1, num_stages: 1 | |
[2023-07-06 04:20:56,144] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 1 configs | |
[2023-07-06 04:20:56,144] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 16, num_warps: 1, num_stages: 1 | |
[2023-07-06 04:20:56,156] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 1 configs | |
[2023-07-06 04:20:56,156] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 32, num_warps: 1, num_stages: 1 | |
[2023-07-06 04:20:57,507] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_ubuntu/hj/chjdplcvb6oibk7aigc4ou76dlz6yn4gxfu3dxactvqvnkkqmftz.py | |
[2023-07-06 04:20:57,508] torch._inductor.compile_fx: [DEBUG] disabled cudagraphs because ['non-cuda device in graph'] | |
skipping cudagraphs due to multiple devices | |
[2023-07-06 04:20:57,508] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 9 | |
[2023-07-06 04:20:57,510] torch._dynamo.output_graph: [INFO] Step 2: done compiler function inductor | |
[2023-07-06 04:20:57,520] torch._dynamo.convert_frame.__guards: [DEBUG] GUARDS: | |
___check_obj_id(L['self'], 140431662763168) | |
hasattr(L['position_ids'], '_dynamo_dynamic_indices') == False | |
hasattr(L['hidden_states'], '_dynamo_dynamic_indices') == False | |
hasattr(L['attention_mask'], '_dynamo_dynamic_indices') == False | |
hasattr(L['past_key_value'], '_dynamo_dynamic_indices') == False | |
hasattr(L['valid_past_index'], '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['output_attentions'], 7698112) | |
not ___is_grad_enabled() | |
not ___are_deterministic_algorithms_enabled() | |
utils_device.CURRENT_DEVICE == None | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks.keys()) == set() | |
[2023-07-06 04:20:57,546] torch._dynamo.convert_frame.__recompiles: [DEBUG] ('Recompiling function <resume in forward> in /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py', 'set env var TORCHDYNAMO_REPORT_GUARD_FAILURES=1 to debug further') | |
[2023-07-06 04:20:57,547] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo start tracing <resume in forward> /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:286 | |
[2023-07-06 04:20:57,547] torch._dynamo.variables.builder: [DEBUG] automatic dynamic L['___stack0'][0] size(1) 1 != 1000 | |
[2023-07-06 04:20:57,547] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['___stack0'][0] (1, 1, 16) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None] | |
[2023-07-06 04:20:57,549] torch._dynamo.variables.builder: [DEBUG] automatic dynamic L['residual'] size(1) 1 != 1000 | |
[2023-07-06 04:20:57,549] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['residual'] (1, 1, 16) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None] | |
[2023-07-06 04:20:57,549] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST ___stack0 [] | |
[2023-07-06 04:20:57,549] torch._dynamo.symbolic_convert: [DEBUG] TRACE JUMP_ABSOLUTE 38 [TupleVariable()] | |
[2023-07-06 04:20:57,549] torch._dynamo.symbolic_convert: [DEBUG] TRACE UNPACK_SEQUENCE 3 [TupleVariable()] | |
[2023-07-06 04:20:57,549] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [ConstantVariable(NoneType), ConstantVariable(NoneType), TensorVariable()] | |
[2023-07-06 04:20:57,549] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST self_attn_weights [ConstantVariable(NoneType), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:57,549] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST present_key_value [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:57,549] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:294 | |
[2023-07-06 04:20:57,549] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST residual [] | |
[2023-07-06 04:20:57,549] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:57,549] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:57,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:57,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:297 | |
[2023-07-06 04:20:57,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:57,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST residual [TensorVariable()] | |
[2023-07-06 04:20:57,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:298 | |
[2023-07-06 04:20:57,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR post_attention_layernorm [NNModuleVariable()] | |
[2023-07-06 04:20:57,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [NNModuleVariable()] | |
[2023-07-06 04:20:57,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:57,553] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:57,554] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:57,554] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:57,554] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:57,555] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:57,555] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:57,555] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:57,555] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,555] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:57,555] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable())] | |
[2023-07-06 04:20:57,555] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:57,555] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,555] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:57,556] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,556] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,556] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:57,556] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,556] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,556] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:57,557] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,557] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,557] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:57,557] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,557] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:57,557] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:57,557] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:57,557] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,557] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:57,557] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:57,557] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:57,557] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,557] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:57,558] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:57,558] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:57,558] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,558] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:57,558] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:57,558] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:57,558] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,558] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:57,558] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:57,558] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable())] | |
[2023-07-06 04:20:57,558] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:57,558] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:57,558] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:57,558] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:57,559] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object forward at 0x7fb8cddd7ec0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 83> | |
[2023-07-06 04:20:57,559] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:84 | |
[2023-07-06 04:20:57,559] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:57,559] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR dtype [TensorVariable()] | |
[2023-07-06 04:20:57,559] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST input_dtype [TorchVariable(torch.float16)] | |
[2023-07-06 04:20:57,559] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85 | |
[2023-07-06 04:20:57,559] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:57,559] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable()] | |
[2023-07-06 04:20:57,559] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:57,559] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR float32 [GetAttrVariable(TensorVariable(), to), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:57,560] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), to), ConstantVariable(dtype)] | |
[2023-07-06 04:20:57,561] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR pow [TensorVariable()] | |
[2023-07-06 04:20:57,561] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [GetAttrVariable(TensorVariable(), pow)] | |
[2023-07-06 04:20:57,561] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), pow), ConstantVariable(int)] | |
[2023-07-06 04:20:57,562] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR mean [TensorVariable()] | |
[2023-07-06 04:20:57,563] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [GetAttrVariable(TensorVariable(), mean)] | |
[2023-07-06 04:20:57,563] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST True [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int)] | |
[2023-07-06 04:20:57,563] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('keepdim',) [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int), ConstantVariable(bool)] | |
[2023-07-06 04:20:57,563] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 2 [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int), ConstantVariable(bool), ConstantVariable(tuple)] | |
[2023-07-06 04:20:57,564] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST variance [TensorVariable()] | |
[2023-07-06 04:20:57,564] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86 | |
[2023-07-06 04:20:57,564] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:57,564] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [TensorVariable()] | |
[2023-07-06 04:20:57,564] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR rsqrt [TensorVariable(), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:57,564] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST variance [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:57,564] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:57,564] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR variance_epsilon [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:57,565] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(float)] | |
[2023-07-06 04:20:57,565] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:57,566] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:57,567] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:57,567] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88 | |
[2023-07-06 04:20:57,567] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,567] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR weight [NNModuleVariable()] | |
[2023-07-06 04:20:57,567] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['self'].post_attention_layernorm.weight (16,) [<DimDynamic.STATIC: 2>] [None] | |
[2023-07-06 04:20:57,568] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:57,568] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:57,569] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable()] | |
[2023-07-06 04:20:57,569] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST input_dtype [GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:57,569] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16)] | |
[2023-07-06 04:20:57,570] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:57,570] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object forward at 0x7fb8cddd7ec0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 83> | |
[2023-07-06 04:20:57,570] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:57,570] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:57,570] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:57,570] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:299 | |
[2023-07-06 04:20:57,571] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,571] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR mlp [NNModuleVariable()] | |
[2023-07-06 04:20:57,571] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [NNModuleVariable()] | |
[2023-07-06 04:20:57,571] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:57,571] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:57,573] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:57,573] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:57,573] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:57,573] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:57,573] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:57,573] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:57,574] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,574] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:57,574] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function LlamaMLP.forward at 0x7fb8d0c464d0>, NNModuleVariable())] | |
[2023-07-06 04:20:57,574] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:57,574] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,574] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:57,574] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,574] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,574] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:57,575] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,575] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,575] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:57,575] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,575] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,575] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:57,575] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:57,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:57,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:57,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:57,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:57,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:57,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:57,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:57,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:57,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:57,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:57,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:57,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:57,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:57,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function LlamaMLP.forward at 0x7fb8d0c464d0>, NNModuleVariable())] | |
[2023-07-06 04:20:57,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function LlamaMLP.forward at 0x7fb8d0c464d0>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:57,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function LlamaMLP.forward at 0x7fb8d0c464d0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:57,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function LlamaMLP.forward at 0x7fb8d0c464d0>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:57,576] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function LlamaMLP.forward at 0x7fb8d0c464d0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:57,577] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object forward at 0x7fb8cddf85b0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 161> | |
[2023-07-06 04:20:57,577] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:162 | |
[2023-07-06 04:20:57,577] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,577] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR down_proj [NNModuleVariable()] | |
[2023-07-06 04:20:57,577] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [NNModuleVariable()] | |
[2023-07-06 04:20:57,577] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR act_fn [NNModuleVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:57,578] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [NNModuleVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:57,578] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR gate_proj [NNModuleVariable(), NNModuleVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:57,578] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [NNModuleVariable(), NNModuleVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:57,578] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), NNModuleVariable(), NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:57,581] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:57,581] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:57,582] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:57,582] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:57,583] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:57,583] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:57,583] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:57,583] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:57,583] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,583] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:57,584] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function SiLUActivation.forward at 0x7fb8ccf64280>, NNModuleVariable())] | |
[2023-07-06 04:20:57,584] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:57,584] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,584] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:57,584] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,584] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,584] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:57,584] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,584] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,584] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:57,585] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,585] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,585] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:57,585] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,585] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:57,585] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:57,585] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:57,585] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,585] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:57,585] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:57,585] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:57,585] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,585] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:57,585] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:57,585] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:57,585] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,586] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:57,586] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:57,586] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:57,586] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,586] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:57,586] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:57,586] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function SiLUActivation.forward at 0x7fb8ccf64280>, NNModuleVariable())] | |
[2023-07-06 04:20:57,586] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function SiLUActivation.forward at 0x7fb8ccf64280>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:57,586] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function SiLUActivation.forward at 0x7fb8ccf64280>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:57,586] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function SiLUActivation.forward at 0x7fb8ccf64280>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:57,586] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function SiLUActivation.forward at 0x7fb8ccf64280>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:57,586] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object forward at 0x7fb8ccf62760, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/activations.py", line 149> | |
[2023-07-06 04:20:57,586] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/activations.py:150 | |
[2023-07-06 04:20:57,586] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL nn [] | |
[2023-07-06 04:20:57,587] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR functional [TorchVariable(<module 'torch.nn' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/__init__.py'>)] | |
[2023-07-06 04:20:57,587] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR silu [TorchVariable(<module 'torch.nn.functional' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/functional.py'>)] | |
[2023-07-06 04:20:57,587] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST input [TorchVariable(<function silu at 0x7fb945bc5990>)] | |
[2023-07-06 04:20:57,587] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [TorchVariable(<function silu at 0x7fb945bc5990>), TensorVariable()] | |
[2023-07-06 04:20:57,590] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:57,590] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object forward at 0x7fb8ccf62760, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/activations.py", line 149> | |
[2023-07-06 04:20:57,590] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:57,590] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:57,590] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:57,591] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR up_proj [NNModuleVariable(), TensorVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:57,591] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [NNModuleVariable(), TensorVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:57,591] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable(), NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:57,593] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [NNModuleVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:57,594] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:57,597] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:57,597] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object forward at 0x7fb8cddf85b0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 161> | |
[2023-07-06 04:20:57,597] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:57,597] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:57,597] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:57,597] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:300 | |
[2023-07-06 04:20:57,597] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST residual [] | |
[2023-07-06 04:20:57,597] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:57,597] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:57,598] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:57,598] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:302 | |
[2023-07-06 04:20:57,598] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:57,598] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_TUPLE 1 [TensorVariable()] | |
[2023-07-06 04:20:57,598] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST outputs [TupleVariable()] | |
[2023-07-06 04:20:57,599] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:304 | |
[2023-07-06 04:20:57,599] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST output_attentions [] | |
[2023-07-06 04:20:57,599] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 106 [ConstantVariable(bool)] | |
[2023-07-06 04:20:57,599] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:307 | |
[2023-07-06 04:20:57,599] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST outputs [] | |
[2023-07-06 04:20:57,599] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST present_key_value [TupleVariable()] | |
[2023-07-06 04:20:57,599] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_TUPLE 1 [TupleVariable(), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:57,599] torch._dynamo.symbolic_convert: [DEBUG] TRACE INPLACE_ADD None [TupleVariable(), TupleVariable()] | |
[2023-07-06 04:20:57,599] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST outputs [TupleVariable()] | |
[2023-07-06 04:20:57,599] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:309 | |
[2023-07-06 04:20:57,599] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST outputs [] | |
[2023-07-06 04:20:57,599] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TupleVariable()] | |
[2023-07-06 04:20:57,599] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo done tracing <resume in forward> (RETURN_VALUE) | |
[2023-07-06 04:20:57,599] torch._dynamo.symbolic_convert: [DEBUG] RETURN_VALUE triggered compile | |
[2023-07-06 04:20:57,599] torch._dynamo.output_graph: [DEBUG] COMPILING GRAPH due to GraphCompileReason(reason='return_value', user_stack=[<FrameSummary file /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py, line 309 in <resume in forward>>], graph_break=False) | |
[2023-07-06 04:20:57,601] torch._dynamo.output_graph.__graph_code: [DEBUG] TRACED GRAPH | |
===== __compiled_fn_18 ===== | |
<eval_with_key>.191 class GraphModule(torch.nn.Module): | |
def forward(self, L_stack0_0_ : torch.Tensor, L_residual_ : torch.Tensor): | |
l_stack0_0_ = L_stack0_0_ | |
l_residual_ = L_residual_ | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:294, code: hidden_states = residual + hidden_states | |
add = l_residual_ + l_stack0_0_; l_residual_ = l_stack0_0_ = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85, code: variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) | |
to = add.to(torch.float32) | |
pow_1 = to.pow(2); to = None | |
mean = pow_1.mean(-1, keepdim = True); pow_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86, code: hidden_states = hidden_states * torch.rsqrt(variance + self.variance_epsilon) | |
add_1 = mean + 1e-06; mean = None | |
rsqrt = torch.rsqrt(add_1); add_1 = None | |
mul = add * rsqrt; rsqrt = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88, code: return (self.weight * hidden_states).to(input_dtype) | |
l__self___post_attention_layernorm_weight = self.L__self___post_attention_layernorm_weight | |
mul_1 = l__self___post_attention_layernorm_weight * mul; l__self___post_attention_layernorm_weight = mul = None | |
to_1 = mul_1.to(torch.float16); mul_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:162, code: return self.down_proj(self.act_fn(self.gate_proj(x)) * self.up_proj(x)) | |
l__self___mlp_gate_proj = self.L__self___mlp_gate_proj(to_1) | |
# File: /home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/activations.py:150, code: return nn.functional.silu(input) | |
silu = torch.nn.functional.silu(l__self___mlp_gate_proj); l__self___mlp_gate_proj = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:162, code: return self.down_proj(self.act_fn(self.gate_proj(x)) * self.up_proj(x)) | |
l__self___mlp_up_proj = self.L__self___mlp_up_proj(to_1); to_1 = None | |
mul_2 = silu * l__self___mlp_up_proj; silu = l__self___mlp_up_proj = None | |
l__self___mlp_down_proj = self.L__self___mlp_down_proj(mul_2); mul_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:300, code: hidden_states = residual + hidden_states | |
add_2 = add + l__self___mlp_down_proj; add = l__self___mlp_down_proj = None | |
return (add_2,) | |
[2023-07-06 04:20:57,602] torch._dynamo.output_graph.__graph: [DEBUG] TRACED GRAPH | |
__compiled_fn_18 <eval_with_key>.191 opcode name target args kwargs | |
------------- ----------------------------------------- -------------------------------------------------------- ------------------------------------------------ ----------------- | |
placeholder l_stack0_0_ L_stack0_0_ () {} | |
placeholder l_residual_ L_residual_ () {} | |
call_function add <built-in function add> (l_residual_, l_stack0_0_) {} | |
call_method to to (add, torch.float32) {} | |
call_method pow_1 pow (to, 2) {} | |
call_method mean mean (pow_1, -1) {'keepdim': True} | |
call_function add_1 <built-in function add> (mean, 1e-06) {} | |
call_function rsqrt <built-in method rsqrt of type object at 0x7fb940469580> (add_1,) {} | |
call_function mul <built-in function mul> (add, rsqrt) {} | |
get_attr l__self___post_attention_layernorm_weight L__self___post_attention_layernorm_weight () {} | |
call_function mul_1 <built-in function mul> (l__self___post_attention_layernorm_weight, mul) {} | |
call_method to_1 to (mul_1, torch.float16) {} | |
call_module l__self___mlp_gate_proj L__self___mlp_gate_proj (to_1,) {} | |
call_function silu <function silu at 0x7fb945bc5990> (l__self___mlp_gate_proj,) {} | |
call_module l__self___mlp_up_proj L__self___mlp_up_proj (to_1,) {} | |
call_function mul_2 <built-in function mul> (silu, l__self___mlp_up_proj) {} | |
call_module l__self___mlp_down_proj L__self___mlp_down_proj (mul_2,) {} | |
call_function add_2 <built-in function add> (add, l__self___mlp_down_proj) {} | |
output output output ((add_2,),) {} | |
[2023-07-06 04:20:57,603] torch._dynamo.output_graph: [INFO] Step 2: calling compiler function inductor | |
[2023-07-06 04:20:57,651] torch._functorch.aot_autograd.__aot_graphs: [INFO] TRACED GRAPH | |
===== Forward graph 23 ===== | |
<eval_with_key>.195 class <lambda>(torch.nn.Module): | |
def forward(self, arg0_1: f16[16], arg1_1: f16[64, 16], arg2_1: f16[64, 16], arg3_1: f16[16, 64], arg4_1: f16[1, 1, 16], arg5_1: f16[1, 1, 16]): | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:294, code: hidden_states = residual + hidden_states | |
add: f16[1, 1, 16] = torch.ops.aten.add.Tensor(arg5_1, arg4_1); arg5_1 = arg4_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85, code: variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) | |
convert_element_type: f32[1, 1, 16] = torch.ops.prims.convert_element_type.default(add, torch.float32) | |
pow_1: f32[1, 1, 16] = torch.ops.aten.pow.Tensor_Scalar(convert_element_type, 2); convert_element_type = None | |
mean: f32[1, 1, 1] = torch.ops.aten.mean.dim(pow_1, [-1], True); pow_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86, code: hidden_states = hidden_states * torch.rsqrt(variance + self.variance_epsilon) | |
add_1: f32[1, 1, 1] = torch.ops.aten.add.Tensor(mean, 1e-06); mean = None | |
rsqrt: f32[1, 1, 1] = torch.ops.aten.rsqrt.default(add_1); add_1 = None | |
mul: f32[1, 1, 16] = torch.ops.aten.mul.Tensor(add, rsqrt); rsqrt = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88, code: return (self.weight * hidden_states).to(input_dtype) | |
mul_1: f32[1, 1, 16] = torch.ops.aten.mul.Tensor(arg0_1, mul); arg0_1 = mul = None | |
convert_element_type_1: f16[1, 1, 16] = torch.ops.prims.convert_element_type.default(mul_1, torch.float16); mul_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:162, code: return self.down_proj(self.act_fn(self.gate_proj(x)) * self.up_proj(x)) | |
permute: f16[16, 64] = torch.ops.aten.permute.default(arg1_1, [1, 0]); arg1_1 = None | |
view: f16[1, 16] = torch.ops.aten.view.default(convert_element_type_1, [1, 16]) | |
mm: f16[1, 64] = torch.ops.aten.mm.default(view, permute); view = permute = None | |
view_1: f16[1, 1, 64] = torch.ops.aten.view.default(mm, [1, 1, 64]); mm = None | |
# File: /home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/activations.py:150, code: return nn.functional.silu(input) | |
convert_element_type_2: f32[1, 1, 64] = torch.ops.prims.convert_element_type.default(view_1, torch.float32); view_1 = None | |
sigmoid: f32[1, 1, 64] = torch.ops.aten.sigmoid.default(convert_element_type_2) | |
mul_2: f32[1, 1, 64] = torch.ops.aten.mul.Tensor(convert_element_type_2, sigmoid); convert_element_type_2 = sigmoid = None | |
convert_element_type_3: f16[1, 1, 64] = torch.ops.prims.convert_element_type.default(mul_2, torch.float16); mul_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:162, code: return self.down_proj(self.act_fn(self.gate_proj(x)) * self.up_proj(x)) | |
permute_1: f16[16, 64] = torch.ops.aten.permute.default(arg2_1, [1, 0]); arg2_1 = None | |
view_2: f16[1, 16] = torch.ops.aten.view.default(convert_element_type_1, [1, 16]); convert_element_type_1 = None | |
mm_1: f16[1, 64] = torch.ops.aten.mm.default(view_2, permute_1); view_2 = permute_1 = None | |
view_3: f16[1, 1, 64] = torch.ops.aten.view.default(mm_1, [1, 1, 64]); mm_1 = None | |
mul_3: f16[1, 1, 64] = torch.ops.aten.mul.Tensor(convert_element_type_3, view_3); convert_element_type_3 = view_3 = None | |
permute_2: f16[64, 16] = torch.ops.aten.permute.default(arg3_1, [1, 0]); arg3_1 = None | |
view_4: f16[1, 64] = torch.ops.aten.view.default(mul_3, [1, 64]); mul_3 = None | |
mm_2: f16[1, 16] = torch.ops.aten.mm.default(view_4, permute_2); view_4 = permute_2 = None | |
view_5: f16[1, 1, 16] = torch.ops.aten.view.default(mm_2, [1, 1, 16]); mm_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:300, code: hidden_states = residual + hidden_states | |
add_2: f16[1, 1, 16] = torch.ops.aten.add.Tensor(add, view_5); add = view_5 = None | |
return (add_2,) | |
[2023-07-06 04:20:57,695] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 10 | |
[2023-07-06 04:20:57,737] torch._inductor.graph: [DEBUG] Force channels last inputs for 0 conv for the current graph with id 10 | |
[2023-07-06 04:20:57,754] torch._inductor.scheduler: [DEBUG] remove_buffer('buf0') | |
[2023-07-06 04:20:57,768] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 1 configs | |
[2023-07-06 04:20:57,768] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 1, num_warps: 2, num_stages: 1 | |
[2023-07-06 04:20:57,768] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 1 configs | |
[2023-07-06 04:20:57,768] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 1 configs | |
[2023-07-06 04:20:57,768] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 64, num_warps: 1, num_stages: 1 | |
[2023-07-06 04:20:57,768] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 16, num_warps: 1, num_stages: 1 | |
[2023-07-06 04:20:57,864] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 1 configs | |
[2023-07-06 04:20:57,864] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 1, num_warps: 2, num_stages: 1 | |
[2023-07-06 04:20:57,928] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 1 configs | |
[2023-07-06 04:20:57,928] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 64, num_warps: 1, num_stages: 1 | |
[2023-07-06 04:20:57,934] torch._inductor.triton_heuristics: [DEBUG] CachingAutotuner gets 1 configs | |
[2023-07-06 04:20:57,934] torch._inductor.triton_heuristics: [DEBUG] XBLOCK: 16, num_warps: 1, num_stages: 1 | |
[2023-07-06 04:20:57,938] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_ubuntu/oo/cooxrverhcp2snw6sgvi3xmyupj2eewevtsplvvj674cqwmsnljq.py | |
[2023-07-06 04:20:57,939] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 10 | |
[2023-07-06 04:20:57,941] torch._dynamo.output_graph: [INFO] Step 2: done compiler function inductor | |
[2023-07-06 04:20:57,949] torch._dynamo.convert_frame.__guards: [DEBUG] GUARDS: | |
___check_obj_id(L['self'], 140431662763072) | |
hasattr(L['residual'], '_dynamo_dynamic_indices') == False | |
___check_type_id(L['___stack0'], 7638272) | |
len(L['___stack0']) == 3 | |
hasattr(L['___stack0'][0], '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['___stack0'][1], 7649088) | |
___check_obj_id(L['___stack0'][2], 7649088) | |
___check_obj_id(L['output_attentions'], 7698112) | |
not ___is_grad_enabled() | |
not ___are_deterministic_algorithms_enabled() | |
utils_device.CURRENT_DEVICE == None | |
str(G['torch'].float32) == 'torch.float32' | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks.keys()) == set() | |
[2023-07-06 04:20:57,949] torch._inductor.cudagraph_trees: [DEBUG] Running warmup of function 7 | |
[2023-07-06 04:20:57,950] torch._dynamo.convert_frame.__recompiles: [DEBUG] ('Recompiling function forward in /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py', 'set env var TORCHDYNAMO_REPORT_GUARD_FAILURES=1 to debug further') | |
[2023-07-06 04:20:57,950] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo start tracing forward /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:261 | |
[2023-07-06 04:20:57,951] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['hidden_states'] (1, 1, 16) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None] | |
[2023-07-06 04:20:57,952] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['attention_mask'] (1, 1003) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None] | |
[2023-07-06 04:20:57,953] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['position_ids'] (1, 1) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None] | |
[2023-07-06 04:20:57,954] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['past_key_value'] (2, 1, 4, 1003, 4) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None, None, None] | |
[2023-07-06 04:20:57,954] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['valid_past_index'] () [] [] | |
[2023-07-06 04:20:57,955] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:281 | |
[2023-07-06 04:20:57,955] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:57,955] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST residual [TensorVariable()] | |
[2023-07-06 04:20:57,955] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:283 | |
[2023-07-06 04:20:57,955] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,955] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR input_layernorm [NNModuleVariable()] | |
[2023-07-06 04:20:57,956] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [NNModuleVariable()] | |
[2023-07-06 04:20:57,956] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:57,956] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:57,957] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:57,957] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:57,958] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:57,958] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:57,958] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:57,958] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:57,958] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,958] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:57,959] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable())] | |
[2023-07-06 04:20:57,959] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:57,959] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,959] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:57,959] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,959] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,959] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:57,959] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,960] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,960] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:57,960] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,960] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,960] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:57,960] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,960] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:57,960] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:57,960] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:57,960] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,961] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:57,961] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:57,961] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:57,961] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,961] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:57,961] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:57,961] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:57,961] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,961] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:57,961] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:57,961] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:57,961] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,961] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:57,961] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:57,961] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable())] | |
[2023-07-06 04:20:57,961] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:57,961] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:57,961] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:57,962] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:57,962] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object forward at 0x7fb8cddd7ec0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 83> | |
[2023-07-06 04:20:57,962] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:84 | |
[2023-07-06 04:20:57,962] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:57,962] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR dtype [TensorVariable()] | |
[2023-07-06 04:20:57,965] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST input_dtype [TorchVariable(torch.float16)] | |
[2023-07-06 04:20:57,965] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85 | |
[2023-07-06 04:20:57,965] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:57,965] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable()] | |
[2023-07-06 04:20:57,965] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:57,965] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR float32 [GetAttrVariable(TensorVariable(), to), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:57,966] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), to), ConstantVariable(dtype)] | |
[2023-07-06 04:20:57,967] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR pow [TensorVariable()] | |
[2023-07-06 04:20:57,967] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [GetAttrVariable(TensorVariable(), pow)] | |
[2023-07-06 04:20:57,967] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), pow), ConstantVariable(int)] | |
[2023-07-06 04:20:57,968] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR mean [TensorVariable()] | |
[2023-07-06 04:20:57,968] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [GetAttrVariable(TensorVariable(), mean)] | |
[2023-07-06 04:20:57,968] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST True [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int)] | |
[2023-07-06 04:20:57,968] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('keepdim',) [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int), ConstantVariable(bool)] | |
[2023-07-06 04:20:57,969] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 2 [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int), ConstantVariable(bool), ConstantVariable(tuple)] | |
[2023-07-06 04:20:57,969] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST variance [TensorVariable()] | |
[2023-07-06 04:20:57,970] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86 | |
[2023-07-06 04:20:57,970] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:57,970] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [TensorVariable()] | |
[2023-07-06 04:20:57,970] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR rsqrt [TensorVariable(), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:57,970] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST variance [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:57,970] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:57,970] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR variance_epsilon [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:57,970] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(float)] | |
[2023-07-06 04:20:57,971] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:57,972] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:57,973] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:57,973] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88 | |
[2023-07-06 04:20:57,973] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,973] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR weight [NNModuleVariable()] | |
[2023-07-06 04:20:57,973] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['self'].input_layernorm.weight (16,) [<DimDynamic.STATIC: 2>] [None] | |
[2023-07-06 04:20:57,974] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:57,974] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:57,975] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable()] | |
[2023-07-06 04:20:57,975] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST input_dtype [GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:57,975] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16)] | |
[2023-07-06 04:20:57,976] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:57,976] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object forward at 0x7fb8cddd7ec0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 83> | |
[2023-07-06 04:20:57,976] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:57,976] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:57,976] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:57,976] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:286 | |
[2023-07-06 04:20:57,976] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,976] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR self_attn [NNModuleVariable()] | |
[2023-07-06 04:20:57,977] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:287 | |
[2023-07-06 04:20:57,977] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [NNModuleVariable()] | |
[2023-07-06 04:20:57,977] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:288 | |
[2023-07-06 04:20:57,977] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST attention_mask [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:57,977] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:289 | |
[2023-07-06 04:20:57,977] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST position_ids [NNModuleVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:57,977] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:290 | |
[2023-07-06 04:20:57,977] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST past_key_value [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:57,977] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:291 | |
[2023-07-06 04:20:57,977] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST output_attentions [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:57,977] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:292 | |
[2023-07-06 04:20:57,977] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ConstantVariable(bool)] | |
[2023-07-06 04:20:57,977] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:286 | |
[2023-07-06 04:20:57,977] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('hidden_states', 'attention_mask', 'position_ids', 'past_key_value', 'output_attentions', 'valid_past_index') [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ConstantVariable(bool), TensorVariable()] | |
[2023-07-06 04:20:57,977] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 6 [NNModuleVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable(), ConstantVariable(bool), TensorVariable(), ConstantVariable(tuple)] | |
[2023-07-06 04:20:57,978] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:57,979] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:57,979] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:57,979] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:57,979] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:57,980] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:57,980] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:57,980] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,980] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:57,980] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function LlamaAttention.forward at 0x7fb8d0c46680>, NNModuleVariable())] | |
[2023-07-06 04:20:57,980] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:57,980] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,980] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:57,981] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,981] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,981] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:57,981] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,981] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,981] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:57,981] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,981] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:57,981] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:57,982] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,982] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:57,982] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:57,982] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:57,982] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,982] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:57,982] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:57,982] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:57,982] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,982] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:57,982] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:57,982] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:57,982] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,982] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:57,982] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:57,982] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:57,982] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:57,982] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:57,982] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:57,982] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function LlamaAttention.forward at 0x7fb8d0c46680>, NNModuleVariable())] | |
[2023-07-06 04:20:57,982] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function LlamaAttention.forward at 0x7fb8d0c46680>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:57,982] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function LlamaAttention.forward at 0x7fb8d0c46680>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:57,983] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function LlamaAttention.forward at 0x7fb8d0c46680>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:57,983] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function LlamaAttention.forward at 0x7fb8d0c46680>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:57,983] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:57,983] torch._dynamo.symbolic_convert: [DEBUG] FAILED INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:57,983] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:57,983] torch._dynamo.symbolic_convert: [DEBUG] break_graph_if_unsupported triggered compile | |
Traceback (most recent call last): | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2187, in inline_call_ | |
sub_locals, closure_cells = func.bind_args(parent, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 187, in bind_args | |
[ | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 188, in <listcomp> | |
wrap(val=arg, source=source) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 62, in wrap_bound_arg | |
return VariableBuilder(tx, source=source, **options)(val) | |
TypeError: VariableBuilder.__init__() got an unexpected keyword argument 'guards' | |
During handling of the above exception, another exception occurred: | |
Traceback (most recent call last): | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 390, in wrapper | |
return inner_fn(self, inst) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 1152, in CALL_FUNCTION_KW | |
self.call_function(fn, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 559, in call_function | |
self.push(fn.call_function(self, args, kwargs)) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/nn_module.py", line 333, in call_function | |
return tx.inline_user_function_return( | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 595, in inline_user_function_return | |
result = InliningInstructionTranslator.inline_call(self, fn, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2132, in inline_call | |
return cls.inline_call_(parent, func, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2229, in inline_call_ | |
tracer.run() | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 708, in run | |
and self.step() | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 668, in step | |
getattr(self, inst.opname)(inst) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 390, in wrapper | |
return inner_fn(self, inst) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 1140, in CALL_FUNCTION_EX | |
self.call_function(fn, argsvars.items, kwargsvars.items) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 559, in call_function | |
self.push(fn.call_function(self, args, kwargs)) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 333, in call_function | |
return super().call_function(tx, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 287, in call_function | |
return super().call_function(tx, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/functions.py", line 120, in call_function | |
return tx.inline_user_function_return( | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 595, in inline_user_function_return | |
result = InliningInstructionTranslator.inline_call(self, fn, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2132, in inline_call | |
return cls.inline_call_(parent, func, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 2190, in inline_call_ | |
raise ArgsMismatchError( | |
torch._dynamo.exc.ArgsMismatchError: VariableBuilder.__init__() got an unexpected keyword argument 'guards'. | |
func = 'forward' /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:190, args = [<class 'trfs_fast.llama.LlamaAttention'>], kwargs = OrderedDict([('hidden_states', TensorVariable()), ('attention_mask', TensorVariable()), ('position_ids', TensorVariable()), ('past_key_value', TensorVariable()), ('output_attentions', ConstantVariable(bool)), ('valid_past_index', TensorVariable())]) | |
[2023-07-06 04:20:57,984] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:57,984] torch._dynamo.output_graph: [DEBUG] COMPILING GRAPH due to GraphCompileReason(reason="VariableBuilder.__init__() got an unexpected keyword argument 'guards'.\n func = 'forward' /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:190, args = [<class 'trfs_fast.llama.LlamaAttention'>], kwargs = OrderedDict([('hidden_states', TensorVariable()), ('attention_mask', TensorVariable()), ('position_ids', TensorVariable()), ('past_key_value', TensorVariable()), ('output_attentions', ConstantVariable(bool)), ('valid_past_index', TensorVariable())])", user_stack=[<FrameSummary file /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py, line 286 in forward>, <FrameSummary file /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py, line 1514 in _call_impl>], graph_break=True) | |
[2023-07-06 04:20:57,985] torch._dynamo.output_graph: [DEBUG] REMOVE UNUSED GRAPHARG L['attention_mask'] | |
[2023-07-06 04:20:57,985] torch._dynamo.output_graph: [DEBUG] REMOVE UNUSED GRAPHARG L['position_ids'] | |
[2023-07-06 04:20:57,985] torch._dynamo.output_graph: [DEBUG] REMOVE UNUSED GRAPHARG L['past_key_value'] | |
[2023-07-06 04:20:57,985] torch._dynamo.output_graph: [DEBUG] REMOVE UNUSED GRAPHARG L['valid_past_index'] | |
[2023-07-06 04:20:57,986] torch._dynamo.output_graph.__graph_code: [DEBUG] TRACED GRAPH | |
===== __compiled_fn_19 ===== | |
<eval_with_key>.204 class GraphModule(torch.nn.Module): | |
def forward(self, L_hidden_states_ : torch.Tensor): | |
l_hidden_states_ = L_hidden_states_ | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85, code: variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) | |
to = l_hidden_states_.to(torch.float32) | |
pow_1 = to.pow(2); to = None | |
mean = pow_1.mean(-1, keepdim = True); pow_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86, code: hidden_states = hidden_states * torch.rsqrt(variance + self.variance_epsilon) | |
add = mean + 1e-06; mean = None | |
rsqrt = torch.rsqrt(add); add = None | |
mul = l_hidden_states_ * rsqrt; l_hidden_states_ = rsqrt = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88, code: return (self.weight * hidden_states).to(input_dtype) | |
l__self___input_layernorm_weight = self.L__self___input_layernorm_weight | |
mul_1 = l__self___input_layernorm_weight * mul; l__self___input_layernorm_weight = mul = None | |
to_1 = mul_1.to(torch.float16); mul_1 = None | |
return (to_1,) | |
[2023-07-06 04:20:57,986] torch._dynamo.output_graph.__graph: [DEBUG] TRACED GRAPH | |
__compiled_fn_19 <eval_with_key>.204 opcode name target args kwargs | |
------------- -------------------------------- -------------------------------------------------------- --------------------------------------- ----------------- | |
placeholder l_hidden_states_ L_hidden_states_ () {} | |
call_method to to (l_hidden_states_, torch.float32) {} | |
call_method pow_1 pow (to, 2) {} | |
call_method mean mean (pow_1, -1) {'keepdim': True} | |
call_function add <built-in function add> (mean, 1e-06) {} | |
call_function rsqrt <built-in method rsqrt of type object at 0x7fb940469580> (add,) {} | |
call_function mul <built-in function mul> (l_hidden_states_, rsqrt) {} | |
get_attr l__self___input_layernorm_weight L__self___input_layernorm_weight () {} | |
call_function mul_1 <built-in function mul> (l__self___input_layernorm_weight, mul) {} | |
call_method to_1 to (mul_1, torch.float16) {} | |
output output output ((to_1,),) {} | |
[2023-07-06 04:20:57,987] torch._dynamo.output_graph: [INFO] Step 2: calling compiler function inductor | |
[2023-07-06 04:20:58,005] torch._functorch.aot_autograd.__aot_graphs: [INFO] TRACED GRAPH | |
===== Forward graph 24 ===== | |
<eval_with_key>.208 class <lambda>(torch.nn.Module): | |
def forward(self, arg0_1: f16[16], arg1_1: f16[1, 1, 16]): | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85, code: variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) | |
convert_element_type: f32[1, 1, 16] = torch.ops.prims.convert_element_type.default(arg1_1, torch.float32) | |
pow_1: f32[1, 1, 16] = torch.ops.aten.pow.Tensor_Scalar(convert_element_type, 2); convert_element_type = None | |
mean: f32[1, 1, 1] = torch.ops.aten.mean.dim(pow_1, [-1], True); pow_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86, code: hidden_states = hidden_states * torch.rsqrt(variance + self.variance_epsilon) | |
add: f32[1, 1, 1] = torch.ops.aten.add.Tensor(mean, 1e-06); mean = None | |
rsqrt: f32[1, 1, 1] = torch.ops.aten.rsqrt.default(add); add = None | |
mul: f32[1, 1, 16] = torch.ops.aten.mul.Tensor(arg1_1, rsqrt); arg1_1 = rsqrt = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88, code: return (self.weight * hidden_states).to(input_dtype) | |
mul_1: f32[1, 1, 16] = torch.ops.aten.mul.Tensor(arg0_1, mul); arg0_1 = mul = None | |
convert_element_type_1: f16[1, 1, 16] = torch.ops.prims.convert_element_type.default(mul_1, torch.float16); mul_1 = None | |
return (convert_element_type_1,) | |
[2023-07-06 04:20:58,007] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 11 | |
[2023-07-06 04:20:58,023] torch._inductor.graph: [DEBUG] Force channels last inputs for 0 conv for the current graph with id 11 | |
[2023-07-06 04:20:58,033] torch._inductor.scheduler: [DEBUG] remove_buffer('buf0') | |
[2023-07-06 04:20:58,034] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_ubuntu/vz/cvzk75oefn2tjj3ekdj3gg5a7rpdjqti5tysf4vatbmcjiobtp5i.py | |
[2023-07-06 04:20:58,035] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 11 | |
[2023-07-06 04:20:58,037] torch._dynamo.output_graph: [INFO] Step 2: done compiler function inductor | |
[2023-07-06 04:20:58,045] torch._dynamo.convert_frame.__guards: [DEBUG] GUARDS: | |
___check_obj_id(L['self'], 140431662486464) | |
hasattr(L['position_ids'], '_dynamo_dynamic_indices') == False | |
___check_type_id(L['hidden_states'], 77765680) | |
hasattr(L['hidden_states'], '_dynamo_dynamic_indices') == False | |
hasattr(L['attention_mask'], '_dynamo_dynamic_indices') == False | |
hasattr(L['past_key_value'], '_dynamo_dynamic_indices') == False | |
hasattr(L['valid_past_index'], '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['output_attentions'], 7698112) | |
not ___is_grad_enabled() | |
not ___are_deterministic_algorithms_enabled() | |
utils_device.CURRENT_DEVICE == None | |
str(G['torch'].float32) == 'torch.float32' | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks.keys()) == set() | |
[2023-07-06 04:20:58,045] torch._inductor.cudagraph_trees: [DEBUG] Running warmup of function 8 | |
[2023-07-06 04:20:58,045] torch._dynamo.convert_frame.__recompiles: [DEBUG] ('Recompiling function forward in /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py', 'set env var TORCHDYNAMO_REPORT_GUARD_FAILURES=1 to debug further') | |
[2023-07-06 04:20:58,046] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo start tracing forward /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:190 | |
[2023-07-06 04:20:58,047] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['hidden_states'] (1, 1, 16) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None] | |
[2023-07-06 04:20:58,048] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['attention_mask'] (1, 1003) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None] | |
[2023-07-06 04:20:58,048] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['position_ids'] (1, 1) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None] | |
[2023-07-06 04:20:58,049] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['past_key_value'] (2, 1, 4, 1003, 4) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None, None, None] | |
[2023-07-06 04:20:58,050] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['valid_past_index'] () [] [] | |
[2023-07-06 04:20:58,051] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:199 | |
[2023-07-06 04:20:58,051] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST output_attentions [] | |
[2023-07-06 04:20:58,051] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST True [ConstantVariable(bool)] | |
[2023-07-06 04:20:58,051] torch._dynamo.symbolic_convert: [DEBUG] TRACE IS_OP 0 [ConstantVariable(bool), ConstantVariable(bool)] | |
[2023-07-06 04:20:58,051] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(bool)] | |
[2023-07-06 04:20:58,051] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:202 | |
[2023-07-06 04:20:58,051] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:58,051] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR size [TensorVariable()] | |
[2023-07-06 04:20:58,051] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [GetAttrVariable(TensorVariable(), size)] | |
[2023-07-06 04:20:58,052] torch._dynamo.symbolic_convert: [DEBUG] TRACE UNPACK_SEQUENCE 3 [SizeVariable()] | |
[2023-07-06 04:20:58,052] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST bsz [ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:58,052] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST q_len [ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:58,052] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST _ [ConstantVariable(int)] | |
[2023-07-06 04:20:58,052] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:205 | |
[2023-07-06 04:20:58,052] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,052] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR qkv_proj [NNModuleVariable()] | |
[2023-07-06 04:20:58,052] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [NNModuleVariable()] | |
[2023-07-06 04:20:58,052] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,055] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR view [TensorVariable()] | |
[2023-07-06 04:20:58,055] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST bsz [GetAttrVariable(TensorVariable(), view)] | |
[2023-07-06 04:20:58,055] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST q_len [GetAttrVariable(TensorVariable(), view), ConstantVariable(int)] | |
[2023-07-06 04:20:58,055] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 3 [GetAttrVariable(TensorVariable(), view), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:58,055] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [GetAttrVariable(TensorVariable(), view), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:58,055] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR num_heads [GetAttrVariable(TensorVariable(), view), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), NNModuleVariable()] | |
[2023-07-06 04:20:58,056] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [GetAttrVariable(TensorVariable(), view), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:58,056] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR head_dim [GetAttrVariable(TensorVariable(), view), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), NNModuleVariable()] | |
[2023-07-06 04:20:58,056] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 5 [GetAttrVariable(TensorVariable(), view), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:58,057] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR permute [TensorVariable()] | |
[2023-07-06 04:20:58,057] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [GetAttrVariable(TensorVariable(), permute)] | |
[2023-07-06 04:20:58,057] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [GetAttrVariable(TensorVariable(), permute), ConstantVariable(int)] | |
[2023-07-06 04:20:58,057] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 3 [GetAttrVariable(TensorVariable(), permute), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:58,057] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [GetAttrVariable(TensorVariable(), permute), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:58,057] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 4 [GetAttrVariable(TensorVariable(), permute), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:58,057] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 5 [GetAttrVariable(TensorVariable(), permute), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:58,058] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST query_key_value_states [TensorVariable()] | |
[2023-07-06 04:20:58,058] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:207 | |
[2023-07-06 04:20:58,058] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST query_key_value_states [] | |
[2023-07-06 04:20:58,058] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [TensorVariable()] | |
[2023-07-06 04:20:58,058] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:58,059] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST query_states [TensorVariable()] | |
[2023-07-06 04:20:58,059] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:208 | |
[2023-07-06 04:20:58,059] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST query_key_value_states [] | |
[2023-07-06 04:20:58,059] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [TensorVariable()] | |
[2023-07-06 04:20:58,059] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:58,059] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_SLICE 2 [TensorVariable(), ConstantVariable(int), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:58,059] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), SliceVariable()] | |
[2023-07-06 04:20:58,061] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST key_value_states [TensorVariable()] | |
[2023-07-06 04:20:58,061] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:211 | |
[2023-07-06 04:20:58,061] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,061] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR rotary_emb [NNModuleVariable()] | |
[2023-07-06 04:20:58,061] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_value_states [NNModuleVariable()] | |
[2023-07-06 04:20:58,061] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,061] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST q_len [NNModuleVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,061] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [NNModuleVariable(), TensorVariable(), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:58,062] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('seq_len',) [NNModuleVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,062] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 2 [NNModuleVariable(), TensorVariable(), TensorVariable(), ConstantVariable(tuple)] | |
[2023-07-06 04:20:58,062] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:58,064] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:58,064] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:58,064] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:58,064] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:58,064] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:58,065] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:58,065] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,065] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:58,065] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function LlamaRotaryEmbedding.forward at 0x7fb8d0c46290>, NNModuleVariable())] | |
[2023-07-06 04:20:58,065] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:58,065] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,065] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:58,065] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,065] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,065] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:58,066] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,066] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,066] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:58,066] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,066] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,066] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:58,067] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,067] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:58,067] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:58,067] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:58,067] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,067] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:58,067] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:58,067] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:58,067] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,067] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:58,067] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:58,067] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:58,067] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,067] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:58,067] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:58,067] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:58,067] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,067] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:58,068] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:58,068] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function LlamaRotaryEmbedding.forward at 0x7fb8d0c46290>, NNModuleVariable())] | |
[2023-07-06 04:20:58,068] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function LlamaRotaryEmbedding.forward at 0x7fb8d0c46290>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:58,068] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function LlamaRotaryEmbedding.forward at 0x7fb8d0c46290>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:58,068] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function LlamaRotaryEmbedding.forward at 0x7fb8d0c46290>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:58,068] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function LlamaRotaryEmbedding.forward at 0x7fb8d0c46290>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:58,068] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object forward at 0x7fb8cddf80e0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 107> | |
[2023-07-06 04:20:58,068] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:122 | |
[2023-07-06 04:20:58,068] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,068] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR cos_cached [NNModuleVariable()] | |
[2023-07-06 04:20:58,069] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['self'].rotary_emb.cos_cached (1, 1, 2048, 4) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None, None] | |
[2023-07-06 04:20:58,069] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable()] | |
[2023-07-06 04:20:58,070] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:58,070] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR dtype [GetAttrVariable(TensorVariable(), to), TensorVariable()] | |
[2023-07-06 04:20:58,070] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('dtype',) [GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16)] | |
[2023-07-06 04:20:58,070] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 1 [GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16), ConstantVariable(tuple)] | |
[2023-07-06 04:20:58,071] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:123 | |
[2023-07-06 04:20:58,071] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [TensorVariable()] | |
[2023-07-06 04:20:58,071] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR sin_cached [TensorVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:58,072] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['self'].rotary_emb.sin_cached (1, 1, 2048, 4) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None, None] | |
[2023-07-06 04:20:58,072] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,072] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TensorVariable(), GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:58,072] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR dtype [TensorVariable(), GetAttrVariable(TensorVariable(), to), TensorVariable()] | |
[2023-07-06 04:20:58,073] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('dtype',) [TensorVariable(), GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16)] | |
[2023-07-06 04:20:58,073] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 1 [TensorVariable(), GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16), ConstantVariable(tuple)] | |
[2023-07-06 04:20:58,074] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:121 | |
[2023-07-06 04:20:58,074] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_TUPLE 2 [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,074] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TupleVariable()] | |
[2023-07-06 04:20:58,074] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object forward at 0x7fb8cddf80e0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 107> | |
[2023-07-06 04:20:58,074] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TupleVariable()] | |
[2023-07-06 04:20:58,074] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:58,074] torch._dynamo.symbolic_convert: [DEBUG] TRACE UNPACK_SEQUENCE 2 [TupleVariable()] | |
[2023-07-06 04:20:58,074] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST cos [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,074] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST sin [TensorVariable()] | |
[2023-07-06 04:20:58,074] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:212 | |
[2023-07-06 04:20:58,074] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL apply_rotary_pos_emb_opt [] | |
[2023-07-06 04:20:58,075] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST query_states [UserFunctionVariable()] | |
[2023-07-06 04:20:58,075] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_value_states [UserFunctionVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,075] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [UserFunctionVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,075] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [UserFunctionVariable(), TensorVariable(), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:58,075] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST cos [UserFunctionVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,075] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST sin [UserFunctionVariable(), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,075] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST position_ids [UserFunctionVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,075] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 5 [UserFunctionVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,076] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object apply_rotary_pos_emb_opt at 0x7fb8cddf82f0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 135> | |
[2023-07-06 04:20:58,076] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:139 | |
[2023-07-06 04:20:58,076] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST cos [] | |
[2023-07-06 04:20:58,076] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR squeeze [TensorVariable()] | |
[2023-07-06 04:20:58,076] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [GetAttrVariable(TensorVariable(), squeeze)] | |
[2023-07-06 04:20:58,076] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), squeeze), ConstantVariable(int)] | |
[2023-07-06 04:20:58,077] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR squeeze [TensorVariable()] | |
[2023-07-06 04:20:58,077] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [GetAttrVariable(TensorVariable(), squeeze)] | |
[2023-07-06 04:20:58,077] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), squeeze), ConstantVariable(int)] | |
[2023-07-06 04:20:58,078] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST cos [TensorVariable()] | |
[2023-07-06 04:20:58,078] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:140 | |
[2023-07-06 04:20:58,078] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST sin [] | |
[2023-07-06 04:20:58,078] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR squeeze [TensorVariable()] | |
[2023-07-06 04:20:58,078] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [GetAttrVariable(TensorVariable(), squeeze)] | |
[2023-07-06 04:20:58,078] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), squeeze), ConstantVariable(int)] | |
[2023-07-06 04:20:58,079] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR squeeze [TensorVariable()] | |
[2023-07-06 04:20:58,079] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [GetAttrVariable(TensorVariable(), squeeze)] | |
[2023-07-06 04:20:58,079] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), squeeze), ConstantVariable(int)] | |
[2023-07-06 04:20:58,080] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST sin [TensorVariable()] | |
[2023-07-06 04:20:58,080] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:141 | |
[2023-07-06 04:20:58,080] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST cos [] | |
[2023-07-06 04:20:58,080] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST position_ids [TensorVariable()] | |
[2023-07-06 04:20:58,080] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,081] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR unsqueeze [TensorVariable()] | |
[2023-07-06 04:20:58,082] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [GetAttrVariable(TensorVariable(), unsqueeze)] | |
[2023-07-06 04:20:58,082] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), unsqueeze), ConstantVariable(int)] | |
[2023-07-06 04:20:58,082] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST cos [TensorVariable()] | |
[2023-07-06 04:20:58,082] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:142 | |
[2023-07-06 04:20:58,082] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST sin [] | |
[2023-07-06 04:20:58,082] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST position_ids [TensorVariable()] | |
[2023-07-06 04:20:58,083] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,084] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR unsqueeze [TensorVariable()] | |
[2023-07-06 04:20:58,084] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [GetAttrVariable(TensorVariable(), unsqueeze)] | |
[2023-07-06 04:20:58,084] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), unsqueeze), ConstantVariable(int)] | |
[2023-07-06 04:20:58,085] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST sin [TensorVariable()] | |
[2023-07-06 04:20:58,085] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:143 | |
[2023-07-06 04:20:58,085] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST q [] | |
[2023-07-06 04:20:58,085] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST cos [TensorVariable()] | |
[2023-07-06 04:20:58,085] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,086] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL rotate_half [TensorVariable()] | |
[2023-07-06 04:20:58,086] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST q [TensorVariable(), UserFunctionVariable()] | |
[2023-07-06 04:20:58,086] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [TensorVariable(), UserFunctionVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,086] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object rotate_half at 0x7fb8cddf8240, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 127> | |
[2023-07-06 04:20:58,086] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:129 | |
[2023-07-06 04:20:58,087] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:58,087] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR roll [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:58,087] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TorchVariable(<built-in method roll of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:58,087] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:58,087] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR shape [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,087] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ShapeVariable()] | |
[2023-07-06 04:20:58,087] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ShapeVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:58,088] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:58,088] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_FLOOR_DIVIDE None [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:58,088] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:58,088] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('shifts', 'dims') [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:58,088] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 3 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int), ConstantVariable(int), ConstantVariable(tuple)] | |
[2023-07-06 04:20:58,089] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST x [TensorVariable()] | |
[2023-07-06 04:20:58,089] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:130 | |
[2023-07-06 04:20:58,089] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [] | |
[2023-07-06 04:20:58,089] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST Ellipsis [TensorVariable()] | |
[2023-07-06 04:20:58,089] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [TensorVariable(), ConstantVariable(ellipsis)] | |
[2023-07-06 04:20:58,089] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:58,089] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR shape [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), TensorVariable()] | |
[2023-07-06 04:20:58,090] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ShapeVariable()] | |
[2023-07-06 04:20:58,090] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ShapeVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:58,090] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ConstantVariable(int)] | |
[2023-07-06 04:20:58,090] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_FLOOR_DIVIDE None [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:58,090] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_SLICE 2 [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ConstantVariable(int)] | |
[2023-07-06 04:20:58,090] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_TUPLE 2 [TensorVariable(), ConstantVariable(ellipsis), SliceVariable()] | |
[2023-07-06 04:20:58,090] torch._dynamo.symbolic_convert: [DEBUG] TRACE DUP_TOP_TWO None [TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:58,090] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), TupleVariable(), TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:58,091] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TensorVariable(), TupleVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,091] torch._dynamo.symbolic_convert: [DEBUG] TRACE INPLACE_MULTIPLY None [TensorVariable(), TupleVariable(), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:58,092] torch._dynamo.symbolic_convert: [DEBUG] TRACE ROT_THREE None [TensorVariable(), TupleVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,092] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_SUBSCR None [TensorVariable(), TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:58,092] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:131 | |
[2023-07-06 04:20:58,092] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [] | |
[2023-07-06 04:20:58,092] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:58,092] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object rotate_half at 0x7fb8cddf8240, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 127> | |
[2023-07-06 04:20:58,092] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST sin [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,092] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,093] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,094] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST q_embed [TensorVariable()] | |
[2023-07-06 04:20:58,094] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:144 | |
[2023-07-06 04:20:58,094] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_states [] | |
[2023-07-06 04:20:58,094] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR copy_ [TensorVariable()] | |
[2023-07-06 04:20:58,094] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_states [GetAttrVariable(TensorVariable(), copy_)] | |
[2023-07-06 04:20:58,094] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST cos [GetAttrVariable(TensorVariable(), copy_), TensorVariable()] | |
[2023-07-06 04:20:58,094] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [GetAttrVariable(TensorVariable(), copy_), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,095] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL rotate_half [GetAttrVariable(TensorVariable(), copy_), TensorVariable()] | |
[2023-07-06 04:20:58,095] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_states [GetAttrVariable(TensorVariable(), copy_), TensorVariable(), UserFunctionVariable()] | |
[2023-07-06 04:20:58,095] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), copy_), TensorVariable(), UserFunctionVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,096] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object rotate_half at 0x7fb8cddf8240, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 127> | |
[2023-07-06 04:20:58,096] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:129 | |
[2023-07-06 04:20:58,096] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:58,096] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR roll [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:58,096] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TorchVariable(<built-in method roll of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:58,096] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:58,096] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR shape [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,097] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ShapeVariable()] | |
[2023-07-06 04:20:58,097] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ShapeVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:58,097] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:58,097] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_FLOOR_DIVIDE None [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:58,097] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:58,097] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('shifts', 'dims') [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:58,097] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 3 [TorchVariable(<built-in method roll of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int), ConstantVariable(int), ConstantVariable(tuple)] | |
[2023-07-06 04:20:58,098] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST x [TensorVariable()] | |
[2023-07-06 04:20:58,098] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:130 | |
[2023-07-06 04:20:58,098] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [] | |
[2023-07-06 04:20:58,098] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST Ellipsis [TensorVariable()] | |
[2023-07-06 04:20:58,098] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [TensorVariable(), ConstantVariable(ellipsis)] | |
[2023-07-06 04:20:58,098] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:58,098] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR shape [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), TensorVariable()] | |
[2023-07-06 04:20:58,099] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ShapeVariable()] | |
[2023-07-06 04:20:58,099] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ShapeVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:58,099] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ConstantVariable(int)] | |
[2023-07-06 04:20:58,099] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_FLOOR_DIVIDE None [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ConstantVariable(int), ConstantVariable(int)] | |
[2023-07-06 04:20:58,099] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_SLICE 2 [TensorVariable(), ConstantVariable(ellipsis), ConstantVariable(NoneType), ConstantVariable(int)] | |
[2023-07-06 04:20:58,099] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_TUPLE 2 [TensorVariable(), ConstantVariable(ellipsis), SliceVariable()] | |
[2023-07-06 04:20:58,099] torch._dynamo.symbolic_convert: [DEBUG] TRACE DUP_TOP_TWO None [TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:58,099] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), TupleVariable(), TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:58,100] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [TensorVariable(), TupleVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,100] torch._dynamo.symbolic_convert: [DEBUG] TRACE INPLACE_MULTIPLY None [TensorVariable(), TupleVariable(), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:58,101] torch._dynamo.symbolic_convert: [DEBUG] TRACE ROT_THREE None [TensorVariable(), TupleVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,101] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_SUBSCR None [TensorVariable(), TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:58,101] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:131 | |
[2023-07-06 04:20:58,101] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [] | |
[2023-07-06 04:20:58,101] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:58,101] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object rotate_half at 0x7fb8cddf8240, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 127> | |
[2023-07-06 04:20:58,101] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST sin [GetAttrVariable(TensorVariable(), copy_), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,101] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [GetAttrVariable(TensorVariable(), copy_), TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,102] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [GetAttrVariable(TensorVariable(), copy_), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,103] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), copy_), TensorVariable()] | |
[2023-07-06 04:20:58,104] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_TOP None [TensorVariable()] | |
[2023-07-06 04:20:58,104] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:145 | |
[2023-07-06 04:20:58,104] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST q_embed [] | |
[2023-07-06 04:20:58,104] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:58,104] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object apply_rotary_pos_emb_opt at 0x7fb8cddf82f0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 135> | |
[2023-07-06 04:20:58,104] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST query_states [TensorVariable()] | |
[2023-07-06 04:20:58,104] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:215 | |
[2023-07-06 04:20:58,104] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [] | |
[2023-07-06 04:20:58,104] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST past_kv_slice_start [TensorVariable()] | |
[2023-07-06 04:20:58,104] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:216 | |
[2023-07-06 04:20:58,104] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:58,104] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR eq [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:58,105] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [TorchVariable(<built-in method eq of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:58,105] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [TorchVariable(<built-in method eq of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:58,105] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 2 [TorchVariable(<built-in method eq of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:58,105] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR int [TensorVariable()] | |
[2023-07-06 04:20:58,106] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [GetAttrVariable(TensorVariable(), int)] | |
[2023-07-06 04:20:58,107] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST q_len [TensorVariable()] | |
[2023-07-06 04:20:58,107] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:58,107] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [TensorVariable()] | |
[2023-07-06 04:20:58,108] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR ne [TensorVariable(), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:58,108] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [TensorVariable(), TorchVariable(<built-in method ne of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:58,108] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [TensorVariable(), TorchVariable(<built-in method ne of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:58,108] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 2 [TensorVariable(), TorchVariable(<built-in method ne of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:58,109] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR int [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,109] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TensorVariable(), GetAttrVariable(TensorVariable(), int)] | |
[2023-07-06 04:20:58,110] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,110] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 1 [TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,110] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TensorVariable(), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:58,111] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,112] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,112] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST past_kv_slice_end [TensorVariable()] | |
[2023-07-06 04:20:58,112] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:217 | |
[2023-07-06 04:20:58,112] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:58,113] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR eq [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:58,113] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [TorchVariable(<built-in method eq of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:58,113] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [TorchVariable(<built-in method eq of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:58,113] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 2 [TorchVariable(<built-in method eq of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:58,114] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR int [TensorVariable()] | |
[2023-07-06 04:20:58,114] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [GetAttrVariable(TensorVariable(), int)] | |
[2023-07-06 04:20:58,115] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_value_states [TensorVariable()] | |
[2023-07-06 04:20:58,115] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR shape [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,115] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -2 [TensorVariable(), ShapeVariable()] | |
[2023-07-06 04:20:58,115] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [TensorVariable(), ShapeVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:58,116] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:58,116] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [TensorVariable()] | |
[2023-07-06 04:20:58,116] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR ne [TensorVariable(), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:58,117] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST valid_past_index [TensorVariable(), TorchVariable(<built-in method ne of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:58,117] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [TensorVariable(), TorchVariable(<built-in method ne of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:58,117] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 2 [TensorVariable(), TorchVariable(<built-in method ne of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:58,118] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR int [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,118] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TensorVariable(), GetAttrVariable(TensorVariable(), int)] | |
[2023-07-06 04:20:58,119] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST past_kv_slice_end [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,119] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,120] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,120] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST past_state_slice_end [TensorVariable()] | |
[2023-07-06 04:20:58,120] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:218 | |
[2023-07-06 04:20:58,120] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST key_value_states [] | |
[2023-07-06 04:20:58,121] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST past_key_value [TensorVariable()] | |
[2023-07-06 04:20:58,121] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST Ellipsis [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,121] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST past_kv_slice_start [TensorVariable(), TensorVariable(), ConstantVariable(ellipsis)] | |
[2023-07-06 04:20:58,121] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST past_kv_slice_end [TensorVariable(), TensorVariable(), ConstantVariable(ellipsis), TensorVariable()] | |
[2023-07-06 04:20:58,121] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_SLICE 2 [TensorVariable(), TensorVariable(), ConstantVariable(ellipsis), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,121] torch._dynamo.symbolic_convert: [DEBUG] step triggered compile | |
Traceback (most recent call last): | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 668, in step | |
getattr(self, inst.opname)(inst) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 1251, in BUILD_SLICE | |
SliceVariable( | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/base.py", line 27, in __call__ | |
obj = type.__call__(cls, *args, **kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/lists.py", line 616, in __init__ | |
unimplemented("Dynamic slicing on data-dependent value is not supported") | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/exc.py", line 140, in unimplemented | |
raise Unsupported(msg) | |
torch._dynamo.exc.Unsupported: Dynamic slicing on data-dependent value is not supported | |
[2023-07-06 04:20:58,121] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:58,121] torch._dynamo.output_graph: [DEBUG] COMPILING GRAPH due to GraphCompileReason(reason='step_unsupported', user_stack=[<FrameSummary file /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py, line 218 in forward>], graph_break=True) | |
[2023-07-06 04:20:58,123] torch._dynamo.output_graph: [DEBUG] REMOVE UNUSED GRAPHARG L['attention_mask'] | |
[2023-07-06 04:20:58,123] torch._dynamo.output_graph: [DEBUG] REMOVE UNUSED GRAPHARG L['past_key_value'] | |
[2023-07-06 04:20:58,124] torch._dynamo.output_graph.__graph_code: [DEBUG] TRACED GRAPH | |
===== __compiled_fn_21 ===== | |
<eval_with_key>.210 class GraphModule(torch.nn.Module): | |
def forward(self, L_hidden_states_ : torch.Tensor, L_position_ids_ : torch.Tensor, L_valid_past_index_ : torch.Tensor): | |
l_hidden_states_ = L_hidden_states_ | |
l_position_ids_ = L_position_ids_ | |
l_valid_past_index_ = L_valid_past_index_ | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:205, code: query_key_value_states = self.qkv_proj(hidden_states).view(bsz, q_len, 3, self.num_heads, self.head_dim).permute(2, 0, 3, 1, 4) | |
l__self___qkv_proj = self.L__self___qkv_proj(l_hidden_states_); l_hidden_states_ = None | |
view = l__self___qkv_proj.view(1, 1, 3, 4, 4); l__self___qkv_proj = None | |
permute = view.permute(2, 0, 3, 1, 4); view = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:207, code: query_states = query_key_value_states[0] | |
getitem = permute[0] | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:208, code: key_value_states = query_key_value_states[1:] | |
getitem_1 = permute[slice(1, None, None)]; permute = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:211, code: cos, sin = self.rotary_emb(key_value_states, seq_len=valid_past_index + q_len) | |
add = l_valid_past_index_ + 1 | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:122, code: self.cos_cached.to(dtype=x.dtype), | |
l__self___rotary_emb_cos_cached = self.L__self___rotary_emb_cos_cached | |
to = l__self___rotary_emb_cos_cached.to(dtype = torch.float16); l__self___rotary_emb_cos_cached = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:123, code: self.sin_cached.to(dtype=x.dtype), | |
l__self___rotary_emb_sin_cached = self.L__self___rotary_emb_sin_cached | |
to_1 = l__self___rotary_emb_sin_cached.to(dtype = torch.float16); l__self___rotary_emb_sin_cached = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:212, code: query_states = apply_rotary_pos_emb_opt(query_states, key_value_states[0], cos, sin, position_ids) | |
getitem_2 = getitem_1[0] | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:139, code: cos = cos.squeeze(1).squeeze(0) # [seq_len, dim] | |
squeeze = to.squeeze(1); to = None | |
squeeze_1 = squeeze.squeeze(0); squeeze = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:140, code: sin = sin.squeeze(1).squeeze(0) # [seq_len, dim] | |
squeeze_2 = to_1.squeeze(1); to_1 = None | |
squeeze_3 = squeeze_2.squeeze(0); squeeze_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:141, code: cos = cos[position_ids].unsqueeze(1) # [bs, 1, seq_len, dim] | |
getitem_3 = squeeze_1[l_position_ids_]; squeeze_1 = None | |
unsqueeze = getitem_3.unsqueeze(1); getitem_3 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:142, code: sin = sin[position_ids].unsqueeze(1) # [bs, 1, seq_len, dim] | |
getitem_4 = squeeze_3[l_position_ids_]; squeeze_3 = l_position_ids_ = None | |
unsqueeze_1 = getitem_4.unsqueeze(1); getitem_4 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:143, code: q_embed = (q * cos) + (rotate_half(q) * sin) | |
mul = getitem * unsqueeze | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:129, code: x = torch.roll(x, shifts=x.shape[-1] // 2, dims=-1) | |
roll = torch.roll(getitem, shifts = 2, dims = -1); getitem = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:130, code: x[..., : x.shape[-1] // 2] *= -1 | |
getitem_5 = roll[(Ellipsis, slice(None, 2, None))] | |
getitem_5 *= -1; imul = getitem_5; getitem_5 = None | |
roll[(Ellipsis, slice(None, 2, None))] = imul; setitem = roll; imul = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:143, code: q_embed = (q * cos) + (rotate_half(q) * sin) | |
mul_1 = roll * unsqueeze_1; roll = None | |
add_1 = mul + mul_1; mul = mul_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:144, code: key_states.copy_((key_states * cos) + (rotate_half(key_states) * sin)) | |
mul_2 = getitem_2 * unsqueeze; unsqueeze = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:129, code: x = torch.roll(x, shifts=x.shape[-1] // 2, dims=-1) | |
roll_1 = torch.roll(getitem_2, shifts = 2, dims = -1) | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:130, code: x[..., : x.shape[-1] // 2] *= -1 | |
getitem_6 = roll_1[(Ellipsis, slice(None, 2, None))] | |
getitem_6 *= -1; imul_1 = getitem_6; getitem_6 = None | |
roll_1[(Ellipsis, slice(None, 2, None))] = imul_1; setitem_1 = roll_1; imul_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:144, code: key_states.copy_((key_states * cos) + (rotate_half(key_states) * sin)) | |
mul_3 = roll_1 * unsqueeze_1; roll_1 = unsqueeze_1 = None | |
add_2 = mul_2 + mul_3; mul_2 = mul_3 = None | |
copy_ = getitem_2.copy_(add_2); getitem_2 = add_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:216, code: past_kv_slice_end = torch.eq(valid_past_index, 0).int() * q_len + torch.ne(valid_past_index, 0).int() * (valid_past_index + 1) | |
eq = torch.eq(l_valid_past_index_, 0) | |
int_1 = eq.int(); eq = None | |
mul_4 = int_1 * 1; int_1 = None | |
ne = torch.ne(l_valid_past_index_, 0) | |
int_2 = ne.int(); ne = None | |
add_3 = l_valid_past_index_ + 1 | |
mul_5 = int_2 * add_3; int_2 = add_3 = None | |
add_4 = mul_4 + mul_5; mul_4 = mul_5 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:217, code: past_state_slice_end = torch.eq(valid_past_index, 0).int() * key_value_states.shape[-2] + torch.ne(valid_past_index, 0).int() * (past_kv_slice_end) | |
eq_1 = torch.eq(l_valid_past_index_, 0) | |
int_3 = eq_1.int(); eq_1 = None | |
mul_6 = int_3 * 1; int_3 = None | |
ne_1 = torch.ne(l_valid_past_index_, 0); l_valid_past_index_ = None | |
int_4 = ne_1.int(); ne_1 = None | |
mul_7 = int_4 * add_4; int_4 = None | |
add_5 = mul_6 + mul_7; mul_6 = mul_7 = None | |
return (add_1, getitem_1, add_4, add_5) | |
[2023-07-06 04:20:58,125] torch._dynamo.output_graph.__graph: [DEBUG] TRACED GRAPH | |
__compiled_fn_21 <eval_with_key>.210 opcode name target args kwargs | |
------------- ------------------------------- ------------------------------------------------------- -------------------------------------------------- ------------------------- | |
placeholder l_hidden_states_ L_hidden_states_ () {} | |
placeholder l_position_ids_ L_position_ids_ () {} | |
placeholder l_valid_past_index_ L_valid_past_index_ () {} | |
call_module l__self___qkv_proj L__self___qkv_proj (l_hidden_states_,) {} | |
call_method view view (l__self___qkv_proj, 1, 1, 3, 4, 4) {} | |
call_method permute permute (view, 2, 0, 3, 1, 4) {} | |
call_function getitem <built-in function getitem> (permute, 0) {} | |
call_function getitem_1 <built-in function getitem> (permute, slice(1, None, None)) {} | |
call_function add <built-in function add> (l_valid_past_index_, 1) {} | |
get_attr l__self___rotary_emb_cos_cached L__self___rotary_emb_cos_cached () {} | |
call_method to to (l__self___rotary_emb_cos_cached,) {'dtype': torch.float16} | |
get_attr l__self___rotary_emb_sin_cached L__self___rotary_emb_sin_cached () {} | |
call_method to_1 to (l__self___rotary_emb_sin_cached,) {'dtype': torch.float16} | |
call_function getitem_2 <built-in function getitem> (getitem_1, 0) {} | |
call_method squeeze squeeze (to, 1) {} | |
call_method squeeze_1 squeeze (squeeze, 0) {} | |
call_method squeeze_2 squeeze (to_1, 1) {} | |
call_method squeeze_3 squeeze (squeeze_2, 0) {} | |
call_function getitem_3 <built-in function getitem> (squeeze_1, l_position_ids_) {} | |
call_method unsqueeze unsqueeze (getitem_3, 1) {} | |
call_function getitem_4 <built-in function getitem> (squeeze_3, l_position_ids_) {} | |
call_method unsqueeze_1 unsqueeze (getitem_4, 1) {} | |
call_function mul <built-in function mul> (getitem, unsqueeze) {} | |
call_function roll <built-in method roll of type object at 0x7fb940469580> (getitem,) {'shifts': 2, 'dims': -1} | |
call_function getitem_5 <built-in function getitem> (roll, (Ellipsis, slice(None, 2, None))) {} | |
call_function imul <built-in function imul> (getitem_5, -1) {} | |
call_function setitem <built-in function setitem> (roll, (Ellipsis, slice(None, 2, None)), imul) {} | |
call_function mul_1 <built-in function mul> (roll, unsqueeze_1) {} | |
call_function add_1 <built-in function add> (mul, mul_1) {} | |
call_function mul_2 <built-in function mul> (getitem_2, unsqueeze) {} | |
call_function roll_1 <built-in method roll of type object at 0x7fb940469580> (getitem_2,) {'shifts': 2, 'dims': -1} | |
call_function getitem_6 <built-in function getitem> (roll_1, (Ellipsis, slice(None, 2, None))) {} | |
call_function imul_1 <built-in function imul> (getitem_6, -1) {} | |
call_function setitem_1 <built-in function setitem> (roll_1, (Ellipsis, slice(None, 2, None)), imul_1) {} | |
call_function mul_3 <built-in function mul> (roll_1, unsqueeze_1) {} | |
call_function add_2 <built-in function add> (mul_2, mul_3) {} | |
call_method copy_ copy_ (getitem_2, add_2) {} | |
call_function eq <built-in method eq of type object at 0x7fb940469580> (l_valid_past_index_, 0) {} | |
call_method int_1 int (eq,) {} | |
call_function mul_4 <built-in function mul> (int_1, 1) {} | |
call_function ne <built-in method ne of type object at 0x7fb940469580> (l_valid_past_index_, 0) {} | |
call_method int_2 int (ne,) {} | |
call_function add_3 <built-in function add> (l_valid_past_index_, 1) {} | |
call_function mul_5 <built-in function mul> (int_2, add_3) {} | |
call_function add_4 <built-in function add> (mul_4, mul_5) {} | |
call_function eq_1 <built-in method eq of type object at 0x7fb940469580> (l_valid_past_index_, 0) {} | |
call_method int_3 int (eq_1,) {} | |
call_function mul_6 <built-in function mul> (int_3, 1) {} | |
call_function ne_1 <built-in method ne of type object at 0x7fb940469580> (l_valid_past_index_, 0) {} | |
call_method int_4 int (ne_1,) {} | |
call_function mul_7 <built-in function mul> (int_4, add_4) {} | |
call_function add_5 <built-in function add> (mul_6, mul_7) {} | |
output output output ((add_1, getitem_1, add_4, add_5),) {} | |
[2023-07-06 04:20:58,127] torch._dynamo.output_graph: [INFO] Step 2: calling compiler function inductor | |
[2023-07-06 04:20:58,248] torch._functorch.aot_autograd.__aot_graphs: [INFO] TRACED GRAPH | |
===== Forward graph 25 ===== | |
<eval_with_key>.214 class <lambda>(torch.nn.Module): | |
def forward(self, arg0_1: f16[48, 16], arg1_1: f32[1, 1, 2048, 4], arg2_1: f32[1, 1, 2048, 4], arg3_1: f16[1, 1, 16], arg4_1: i64[1, 1], arg5_1: i64[]): | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:205, code: query_key_value_states = self.qkv_proj(hidden_states).view(bsz, q_len, 3, self.num_heads, self.head_dim).permute(2, 0, 3, 1, 4) | |
permute: f16[16, 48] = torch.ops.aten.permute.default(arg0_1, [1, 0]); arg0_1 = None | |
view: f16[1, 16] = torch.ops.aten.view.default(arg3_1, [1, 16]); arg3_1 = None | |
mm: f16[1, 48] = torch.ops.aten.mm.default(view, permute); view = permute = None | |
view_1: f16[1, 1, 48] = torch.ops.aten.view.default(mm, [1, 1, 48]) | |
view_2: f16[1, 1, 3, 4, 4] = torch.ops.aten.view.default(view_1, [1, 1, 3, 4, 4]); view_1 = None | |
permute_1: f16[3, 1, 4, 1, 4] = torch.ops.aten.permute.default(view_2, [2, 0, 3, 1, 4]); view_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:207, code: query_states = query_key_value_states[0] | |
select: f16[1, 4, 1, 4] = torch.ops.aten.select.int(permute_1, 0, 0) | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:208, code: key_value_states = query_key_value_states[1:] | |
slice_1: f16[2, 1, 4, 1, 4] = torch.ops.aten.slice.Tensor(permute_1, 0, 1, 9223372036854775807); permute_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:122, code: self.cos_cached.to(dtype=x.dtype), | |
convert_element_type: f16[1, 1, 2048, 4] = torch.ops.prims.convert_element_type.default(arg1_1, torch.float16); arg1_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:123, code: self.sin_cached.to(dtype=x.dtype), | |
convert_element_type_1: f16[1, 1, 2048, 4] = torch.ops.prims.convert_element_type.default(arg2_1, torch.float16); arg2_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:212, code: query_states = apply_rotary_pos_emb_opt(query_states, key_value_states[0], cos, sin, position_ids) | |
select_1: f16[1, 4, 1, 4] = torch.ops.aten.select.int(slice_1, 0, 0); slice_1 = None | |
# No stacktrace found for following nodes | |
squeeze: f16[1, 2048, 4] = torch.ops.aten.squeeze.dim(convert_element_type, 1); convert_element_type = None | |
squeeze_1: f16[2048, 4] = torch.ops.aten.squeeze.dim(squeeze, 0); squeeze = None | |
squeeze_2: f16[1, 2048, 4] = torch.ops.aten.squeeze.dim(convert_element_type_1, 1); convert_element_type_1 = None | |
squeeze_3: f16[2048, 4] = torch.ops.aten.squeeze.dim(squeeze_2, 0); squeeze_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:141, code: cos = cos[position_ids].unsqueeze(1) # [bs, 1, seq_len, dim] | |
index: f16[1, 1, 4] = torch.ops.aten.index.Tensor(squeeze_1, [arg4_1]); squeeze_1 = None | |
unsqueeze: f16[1, 1, 1, 4] = torch.ops.aten.unsqueeze.default(index, 1); index = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:142, code: sin = sin[position_ids].unsqueeze(1) # [bs, 1, seq_len, dim] | |
index_1: f16[1, 1, 4] = torch.ops.aten.index.Tensor(squeeze_3, [arg4_1]); squeeze_3 = arg4_1 = None | |
unsqueeze_1: f16[1, 1, 1, 4] = torch.ops.aten.unsqueeze.default(index_1, 1); index_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:143, code: q_embed = (q * cos) + (rotate_half(q) * sin) | |
mul: f16[1, 4, 1, 4] = torch.ops.aten.mul.Tensor(select, unsqueeze) | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:129, code: x = torch.roll(x, shifts=x.shape[-1] // 2, dims=-1) | |
roll: f16[1, 4, 1, 4] = torch.ops.aten.roll.default(select, [2], [-1]); select = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:130, code: x[..., : x.shape[-1] // 2] *= -1 | |
slice_2: f16[1, 4, 1, 2] = torch.ops.aten.slice.Tensor(roll, 3, 0, 2) | |
mul_1: f16[1, 4, 1, 2] = torch.ops.aten.mul.Tensor(slice_2, -1); slice_2 = None | |
slice_scatter: f16[1, 4, 1, 4] = torch.ops.aten.slice_scatter.default(roll, mul_1, 3, 0, 2); roll = mul_1 = None | |
slice_3: f16[1, 4, 1, 2] = torch.ops.aten.slice.Tensor(slice_scatter, 3, 0, 2) | |
slice_scatter_1: f16[1, 4, 1, 4] = torch.ops.aten.slice_scatter.default(slice_scatter, slice_3, 3, 0, 2); slice_scatter = slice_3 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:143, code: q_embed = (q * cos) + (rotate_half(q) * sin) | |
mul_2: f16[1, 4, 1, 4] = torch.ops.aten.mul.Tensor(slice_scatter_1, unsqueeze_1); slice_scatter_1 = None | |
add_1: f16[1, 4, 1, 4] = torch.ops.aten.add.Tensor(mul, mul_2); mul = mul_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:144, code: key_states.copy_((key_states * cos) + (rotate_half(key_states) * sin)) | |
mul_3: f16[1, 4, 1, 4] = torch.ops.aten.mul.Tensor(select_1, unsqueeze); unsqueeze = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:129, code: x = torch.roll(x, shifts=x.shape[-1] // 2, dims=-1) | |
roll_1: f16[1, 4, 1, 4] = torch.ops.aten.roll.default(select_1, [2], [-1]); select_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:130, code: x[..., : x.shape[-1] // 2] *= -1 | |
slice_7: f16[1, 4, 1, 2] = torch.ops.aten.slice.Tensor(roll_1, 3, 0, 2) | |
mul_4: f16[1, 4, 1, 2] = torch.ops.aten.mul.Tensor(slice_7, -1); slice_7 = None | |
slice_scatter_2: f16[1, 4, 1, 4] = torch.ops.aten.slice_scatter.default(roll_1, mul_4, 3, 0, 2); roll_1 = mul_4 = None | |
slice_8: f16[1, 4, 1, 2] = torch.ops.aten.slice.Tensor(slice_scatter_2, 3, 0, 2) | |
slice_scatter_3: f16[1, 4, 1, 4] = torch.ops.aten.slice_scatter.default(slice_scatter_2, slice_8, 3, 0, 2); slice_scatter_2 = slice_8 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:144, code: key_states.copy_((key_states * cos) + (rotate_half(key_states) * sin)) | |
mul_5: f16[1, 4, 1, 4] = torch.ops.aten.mul.Tensor(slice_scatter_3, unsqueeze_1); slice_scatter_3 = unsqueeze_1 = None | |
add_2: f16[1, 4, 1, 4] = torch.ops.aten.add.Tensor(mul_3, mul_5); mul_3 = mul_5 = None | |
view_3: f16[1, 1, 48] = torch.ops.aten.view.default(mm, [1, 1, 48]); mm = None | |
view_4: f16[1, 1, 3, 4, 4] = torch.ops.aten.view.default(view_3, [1, 1, 3, 4, 4]); view_3 = None | |
permute_2: f16[3, 1, 4, 1, 4] = torch.ops.aten.permute.default(view_4, [2, 0, 3, 1, 4]); view_4 = None | |
slice_12: f16[2, 1, 4, 1, 4] = torch.ops.aten.slice.Tensor(permute_2, 0, 1, 9223372036854775807) | |
select_scatter: f16[2, 1, 4, 1, 4] = torch.ops.aten.select_scatter.default(slice_12, add_2, 0, 0); slice_12 = add_2 = None | |
slice_scatter_4: f16[3, 1, 4, 1, 4] = torch.ops.aten.slice_scatter.default(permute_2, select_scatter, 0, 1, 9223372036854775807); permute_2 = select_scatter = None | |
permute_3: f16[1, 1, 3, 4, 4] = torch.ops.aten.permute.default(slice_scatter_4, [1, 3, 0, 2, 4]); slice_scatter_4 = None | |
view_5: f16[1, 1, 48] = torch.ops.aten.view.default(permute_3, [1, 1, 48]); permute_3 = None | |
view_6: f16[1, 48] = torch.ops.aten.view.default(view_5, [1, 48]); view_5 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:216, code: past_kv_slice_end = torch.eq(valid_past_index, 0).int() * q_len + torch.ne(valid_past_index, 0).int() * (valid_past_index + 1) | |
eq: b8[] = torch.ops.aten.eq.Scalar(arg5_1, 0) | |
convert_element_type_2: i32[] = torch.ops.prims.convert_element_type.default(eq, torch.int32); eq = None | |
mul_6: i32[] = torch.ops.aten.mul.Tensor(convert_element_type_2, 1); convert_element_type_2 = None | |
ne: b8[] = torch.ops.aten.ne.Scalar(arg5_1, 0) | |
convert_element_type_3: i32[] = torch.ops.prims.convert_element_type.default(ne, torch.int32); ne = None | |
add_3: i64[] = torch.ops.aten.add.Tensor(arg5_1, 1) | |
mul_7: i64[] = torch.ops.aten.mul.Tensor(convert_element_type_3, add_3); convert_element_type_3 = add_3 = None | |
add_4: i64[] = torch.ops.aten.add.Tensor(mul_6, mul_7); mul_6 = mul_7 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:217, code: past_state_slice_end = torch.eq(valid_past_index, 0).int() * key_value_states.shape[-2] + torch.ne(valid_past_index, 0).int() * (past_kv_slice_end) | |
eq_1: b8[] = torch.ops.aten.eq.Scalar(arg5_1, 0) | |
convert_element_type_4: i32[] = torch.ops.prims.convert_element_type.default(eq_1, torch.int32); eq_1 = None | |
mul_8: i32[] = torch.ops.aten.mul.Tensor(convert_element_type_4, 1); convert_element_type_4 = None | |
ne_1: b8[] = torch.ops.aten.ne.Scalar(arg5_1, 0); arg5_1 = None | |
convert_element_type_5: i32[] = torch.ops.prims.convert_element_type.default(ne_1, torch.int32); ne_1 = None | |
mul_9: i64[] = torch.ops.aten.mul.Tensor(convert_element_type_5, add_4); convert_element_type_5 = None | |
add_5: i64[] = torch.ops.aten.add.Tensor(mul_8, mul_9); mul_8 = mul_9 = None | |
# No stacktrace found for following nodes | |
view_9: f16[1, 1, 48] = torch.ops.aten.view.default(view_6, [1, 1, 48]); view_6 = None | |
view_10: f16[1, 1, 3, 4, 4] = torch.ops.aten.view.default(view_9, [1, 1, 3, 4, 4]); view_9 = None | |
permute_5: f16[3, 1, 4, 1, 4] = torch.ops.aten.permute.default(view_10, [2, 0, 3, 1, 4]); view_10 = None | |
slice_14: f16[2, 1, 4, 1, 4] = torch.ops.aten.slice.Tensor(permute_5, 0, 1, 9223372036854775807); permute_5 = None | |
return (add_1, slice_14, add_4, add_5) | |
[2023-07-06 04:20:58,284] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 12 | |
[2023-07-06 04:20:58,391] torch._inductor.graph: [DEBUG] Force channels last inputs for 0 conv for the current graph with id 12 | |
[2023-07-06 04:20:58,487] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_ubuntu/hj/chjdplcvb6oibk7aigc4ou76dlz6yn4gxfu3dxactvqvnkkqmftz.py | |
[2023-07-06 04:20:58,487] torch._inductor.compile_fx: [DEBUG] disabled cudagraphs because ['non-cuda device in graph'] | |
skipping cudagraphs due to multiple devices | |
[2023-07-06 04:20:58,487] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 12 | |
[2023-07-06 04:20:58,490] torch._dynamo.output_graph: [INFO] Step 2: done compiler function inductor | |
[2023-07-06 04:20:58,499] torch._dynamo.convert_frame.__guards: [DEBUG] GUARDS: | |
___check_obj_id(L['self'], 140431662491744) | |
hasattr(L['position_ids'], '_dynamo_dynamic_indices') == False | |
hasattr(L['hidden_states'], '_dynamo_dynamic_indices') == False | |
hasattr(L['attention_mask'], '_dynamo_dynamic_indices') == False | |
hasattr(L['past_key_value'], '_dynamo_dynamic_indices') == False | |
hasattr(L['valid_past_index'], '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['output_attentions'], 7698112) | |
not ___is_grad_enabled() | |
not ___are_deterministic_algorithms_enabled() | |
utils_device.CURRENT_DEVICE == None | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks.keys()) == set() | |
[2023-07-06 04:20:58,500] torch._dynamo.convert_frame.__recompiles: [DEBUG] ('Recompiling function <resume in forward> in /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py', 'set env var TORCHDYNAMO_REPORT_GUARD_FAILURES=1 to debug further') | |
[2023-07-06 04:20:58,500] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo start tracing <resume in forward> /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:286 | |
[2023-07-06 04:20:58,500] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['___stack0'][0] (1, 1, 16) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None] | |
[2023-07-06 04:20:58,501] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['residual'] (1, 1, 16) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None] | |
[2023-07-06 04:20:58,503] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST ___stack0 [] | |
[2023-07-06 04:20:58,503] torch._dynamo.symbolic_convert: [DEBUG] TRACE JUMP_ABSOLUTE 38 [TupleVariable()] | |
[2023-07-06 04:20:58,503] torch._dynamo.symbolic_convert: [DEBUG] TRACE UNPACK_SEQUENCE 3 [TupleVariable()] | |
[2023-07-06 04:20:58,503] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [ConstantVariable(NoneType), ConstantVariable(NoneType), TensorVariable()] | |
[2023-07-06 04:20:58,503] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST self_attn_weights [ConstantVariable(NoneType), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:58,503] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST present_key_value [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:58,503] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:294 | |
[2023-07-06 04:20:58,503] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST residual [] | |
[2023-07-06 04:20:58,503] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:58,503] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,504] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:58,504] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:297 | |
[2023-07-06 04:20:58,504] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:58,504] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST residual [TensorVariable()] | |
[2023-07-06 04:20:58,504] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:298 | |
[2023-07-06 04:20:58,504] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,504] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR post_attention_layernorm [NNModuleVariable()] | |
[2023-07-06 04:20:58,504] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [NNModuleVariable()] | |
[2023-07-06 04:20:58,505] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,505] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:58,506] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:58,507] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:58,507] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:58,507] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:58,507] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:58,507] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:58,507] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,507] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:58,508] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable())] | |
[2023-07-06 04:20:58,508] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:58,508] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,508] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:58,508] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,508] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,508] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:58,509] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,509] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,509] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:58,509] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,509] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,509] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:58,509] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,509] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:58,509] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:58,509] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:58,510] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,510] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:58,510] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:58,510] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:58,510] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,510] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:58,510] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:58,510] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:58,510] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,510] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:58,510] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:58,510] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:58,510] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,510] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:58,510] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:58,510] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable())] | |
[2023-07-06 04:20:58,510] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:58,511] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:58,511] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:58,511] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function LlamaRMSNorm.forward at 0x7fb8d0c460e0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:58,511] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object forward at 0x7fb8cddd7ec0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 83> | |
[2023-07-06 04:20:58,511] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:84 | |
[2023-07-06 04:20:58,511] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:58,511] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR dtype [TensorVariable()] | |
[2023-07-06 04:20:58,511] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST input_dtype [TorchVariable(torch.float16)] | |
[2023-07-06 04:20:58,511] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85 | |
[2023-07-06 04:20:58,511] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:58,512] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable()] | |
[2023-07-06 04:20:58,512] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:58,512] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR float32 [GetAttrVariable(TensorVariable(), to), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:58,512] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), to), ConstantVariable(dtype)] | |
[2023-07-06 04:20:58,513] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR pow [TensorVariable()] | |
[2023-07-06 04:20:58,514] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [GetAttrVariable(TensorVariable(), pow)] | |
[2023-07-06 04:20:58,514] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), pow), ConstantVariable(int)] | |
[2023-07-06 04:20:58,515] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR mean [TensorVariable()] | |
[2023-07-06 04:20:58,515] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [GetAttrVariable(TensorVariable(), mean)] | |
[2023-07-06 04:20:58,515] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST True [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int)] | |
[2023-07-06 04:20:58,515] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('keepdim',) [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int), ConstantVariable(bool)] | |
[2023-07-06 04:20:58,515] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 2 [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int), ConstantVariable(bool), ConstantVariable(tuple)] | |
[2023-07-06 04:20:58,516] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST variance [TensorVariable()] | |
[2023-07-06 04:20:58,516] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86 | |
[2023-07-06 04:20:58,516] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:58,516] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [TensorVariable()] | |
[2023-07-06 04:20:58,516] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR rsqrt [TensorVariable(), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:58,516] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST variance [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:58,516] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:58,516] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR variance_epsilon [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:58,517] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(float)] | |
[2023-07-06 04:20:58,517] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:58,518] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,519] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:58,519] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88 | |
[2023-07-06 04:20:58,519] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,519] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR weight [NNModuleVariable()] | |
[2023-07-06 04:20:58,519] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['self'].post_attention_layernorm.weight (16,) [<DimDynamic.STATIC: 2>] [None] | |
[2023-07-06 04:20:58,520] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:58,520] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,521] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable()] | |
[2023-07-06 04:20:58,521] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST input_dtype [GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:58,521] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16)] | |
[2023-07-06 04:20:58,522] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:58,522] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object forward at 0x7fb8cddd7ec0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 83> | |
[2023-07-06 04:20:58,522] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:58,522] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:58,522] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:58,523] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:299 | |
[2023-07-06 04:20:58,523] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,523] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR mlp [NNModuleVariable()] | |
[2023-07-06 04:20:58,523] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [NNModuleVariable()] | |
[2023-07-06 04:20:58,523] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,523] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:58,525] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:58,525] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:58,525] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:58,525] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:58,525] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:58,525] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:58,525] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,526] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:58,526] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function LlamaMLP.forward at 0x7fb8d0c464d0>, NNModuleVariable())] | |
[2023-07-06 04:20:58,526] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:58,526] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,526] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:58,526] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,526] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,526] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:58,527] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,527] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,527] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:58,527] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,527] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,527] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:58,527] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,527] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:58,527] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:58,527] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:58,527] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,528] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:58,528] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:58,528] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:58,528] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,528] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:58,528] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:58,528] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:58,528] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,528] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:58,528] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:58,528] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:58,528] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,528] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:58,528] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:58,528] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function LlamaMLP.forward at 0x7fb8d0c464d0>, NNModuleVariable())] | |
[2023-07-06 04:20:58,528] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function LlamaMLP.forward at 0x7fb8d0c464d0>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:58,528] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function LlamaMLP.forward at 0x7fb8d0c464d0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:58,528] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function LlamaMLP.forward at 0x7fb8d0c464d0>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:58,528] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function LlamaMLP.forward at 0x7fb8d0c464d0>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:58,529] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object forward at 0x7fb8cddf85b0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 161> | |
[2023-07-06 04:20:58,529] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:162 | |
[2023-07-06 04:20:58,529] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,529] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR down_proj [NNModuleVariable()] | |
[2023-07-06 04:20:58,529] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [NNModuleVariable()] | |
[2023-07-06 04:20:58,529] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR act_fn [NNModuleVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:58,529] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [NNModuleVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:58,529] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR gate_proj [NNModuleVariable(), NNModuleVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:58,530] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [NNModuleVariable(), NNModuleVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:58,530] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), NNModuleVariable(), NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,533] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,533] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:58,534] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1508 | |
[2023-07-06 04:20:58,534] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [] | |
[2023-07-06 04:20:58,534] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _C [TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:58,535] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _get_tracing_state [TorchVariable(<module 'torch._C' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_C.cpython-310-x86_64-linux-gnu.so'>)] | |
[2023-07-06 04:20:58,535] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 0 [TorchVariable(<built-in method _get_tracing_state of PyCapsule object at 0x7fb974963060>)] | |
[2023-07-06 04:20:58,535] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 16 [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:58,535] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,535] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR forward [NNModuleVariable()] | |
[2023-07-06 04:20:58,535] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST forward_call [UserMethodVariable(<function SiLUActivation.forward at 0x7fb8ccf64280>, NNModuleVariable())] | |
[2023-07-06 04:20:58,535] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:58,535] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,535] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:58,536] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,536] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,536] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _backward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:58,536] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,536] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,536] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:58,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR _forward_pre_hooks [NNModuleVariable()] | |
[2023-07-06 04:20:58,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:58,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_pre_hooks [] | |
[2023-07-06 04:20:58,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:58,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1512 | |
[2023-07-06 04:20:58,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_backward_hooks [] | |
[2023-07-06 04:20:58,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:58,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:58,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_hooks [] | |
[2023-07-06 04:20:58,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:58,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1513 | |
[2023-07-06 04:20:58,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL _global_forward_pre_hooks [] | |
[2023-07-06 04:20:58,537] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1511 | |
[2023-07-06 04:20:58,538] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 76 [ConstDictVariable()] | |
[2023-07-06 04:20:58,538] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py:1514 | |
[2023-07-06 04:20:58,538] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST forward_call [] | |
[2023-07-06 04:20:58,538] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST args [UserMethodVariable(<function SiLUActivation.forward at 0x7fb8ccf64280>, NNModuleVariable())] | |
[2023-07-06 04:20:58,538] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_MAP 0 [UserMethodVariable(<function SiLUActivation.forward at 0x7fb8ccf64280>, NNModuleVariable()), TupleVariable()] | |
[2023-07-06 04:20:58,538] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST kwargs [UserMethodVariable(<function SiLUActivation.forward at 0x7fb8ccf64280>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:58,538] torch._dynamo.symbolic_convert: [DEBUG] TRACE DICT_MERGE 1 [UserMethodVariable(<function SiLUActivation.forward at 0x7fb8ccf64280>, NNModuleVariable()), TupleVariable(), ConstDictVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:58,538] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_EX 1 [UserMethodVariable(<function SiLUActivation.forward at 0x7fb8ccf64280>, NNModuleVariable()), TupleVariable(), ConstDictVariable()] | |
[2023-07-06 04:20:58,538] torch._dynamo.symbolic_convert: [DEBUG] INLINING <code object forward at 0x7fb8ccf62760, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/activations.py", line 149> | |
[2023-07-06 04:20:58,538] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/activations.py:150 | |
[2023-07-06 04:20:58,538] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL nn [] | |
[2023-07-06 04:20:58,538] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR functional [TorchVariable(<module 'torch.nn' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/__init__.py'>)] | |
[2023-07-06 04:20:58,539] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR silu [TorchVariable(<module 'torch.nn.functional' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/functional.py'>)] | |
[2023-07-06 04:20:58,539] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST input [TorchVariable(<function silu at 0x7fb945bc5990>)] | |
[2023-07-06 04:20:58,539] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [TorchVariable(<function silu at 0x7fb945bc5990>), TensorVariable()] | |
[2023-07-06 04:20:58,543] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:58,543] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object forward at 0x7fb8ccf62760, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/activations.py", line 149> | |
[2023-07-06 04:20:58,543] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:58,543] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:58,543] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,543] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR up_proj [NNModuleVariable(), TensorVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:58,543] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST x [NNModuleVariable(), TensorVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:58,543] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable(), NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,546] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [NNModuleVariable(), TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,547] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,549] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:58,549] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object forward at 0x7fb8cddf85b0, file "/home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py", line 161> | |
[2023-07-06 04:20:58,549] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:58,549] torch._dynamo.symbolic_convert: [DEBUG] DONE INLINING <code object _call_impl at 0x7fb97409ec30, file "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1507> | |
[2023-07-06 04:20:58,549] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:58,549] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:300 | |
[2023-07-06 04:20:58,549] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST residual [] | |
[2023-07-06 04:20:58,549] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:58,550] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,550] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:58,550] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:302 | |
[2023-07-06 04:20:58,550] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:58,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_TUPLE 1 [TensorVariable()] | |
[2023-07-06 04:20:58,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST outputs [TupleVariable()] | |
[2023-07-06 04:20:58,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:304 | |
[2023-07-06 04:20:58,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST output_attentions [] | |
[2023-07-06 04:20:58,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 106 [ConstantVariable(bool)] | |
[2023-07-06 04:20:58,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:307 | |
[2023-07-06 04:20:58,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST outputs [] | |
[2023-07-06 04:20:58,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST present_key_value [TupleVariable()] | |
[2023-07-06 04:20:58,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE BUILD_TUPLE 1 [TupleVariable(), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:58,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE INPLACE_ADD None [TupleVariable(), TupleVariable()] | |
[2023-07-06 04:20:58,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST outputs [TupleVariable()] | |
[2023-07-06 04:20:58,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:309 | |
[2023-07-06 04:20:58,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST outputs [] | |
[2023-07-06 04:20:58,551] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TupleVariable()] | |
[2023-07-06 04:20:58,551] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo done tracing <resume in forward> (RETURN_VALUE) | |
[2023-07-06 04:20:58,551] torch._dynamo.symbolic_convert: [DEBUG] RETURN_VALUE triggered compile | |
[2023-07-06 04:20:58,551] torch._dynamo.output_graph: [DEBUG] COMPILING GRAPH due to GraphCompileReason(reason='return_value', user_stack=[<FrameSummary file /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py, line 309 in <resume in forward>>], graph_break=False) | |
[2023-07-06 04:20:58,553] torch._dynamo.output_graph.__graph_code: [DEBUG] TRACED GRAPH | |
===== __compiled_fn_22 ===== | |
<eval_with_key>.219 class GraphModule(torch.nn.Module): | |
def forward(self, L_stack0_0_ : torch.Tensor, L_residual_ : torch.Tensor): | |
l_stack0_0_ = L_stack0_0_ | |
l_residual_ = L_residual_ | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:294, code: hidden_states = residual + hidden_states | |
add = l_residual_ + l_stack0_0_; l_residual_ = l_stack0_0_ = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85, code: variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) | |
to = add.to(torch.float32) | |
pow_1 = to.pow(2); to = None | |
mean = pow_1.mean(-1, keepdim = True); pow_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86, code: hidden_states = hidden_states * torch.rsqrt(variance + self.variance_epsilon) | |
add_1 = mean + 1e-06; mean = None | |
rsqrt = torch.rsqrt(add_1); add_1 = None | |
mul = add * rsqrt; rsqrt = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88, code: return (self.weight * hidden_states).to(input_dtype) | |
l__self___post_attention_layernorm_weight = self.L__self___post_attention_layernorm_weight | |
mul_1 = l__self___post_attention_layernorm_weight * mul; l__self___post_attention_layernorm_weight = mul = None | |
to_1 = mul_1.to(torch.float16); mul_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:162, code: return self.down_proj(self.act_fn(self.gate_proj(x)) * self.up_proj(x)) | |
l__self___mlp_gate_proj = self.L__self___mlp_gate_proj(to_1) | |
# File: /home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/activations.py:150, code: return nn.functional.silu(input) | |
silu = torch.nn.functional.silu(l__self___mlp_gate_proj); l__self___mlp_gate_proj = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:162, code: return self.down_proj(self.act_fn(self.gate_proj(x)) * self.up_proj(x)) | |
l__self___mlp_up_proj = self.L__self___mlp_up_proj(to_1); to_1 = None | |
mul_2 = silu * l__self___mlp_up_proj; silu = l__self___mlp_up_proj = None | |
l__self___mlp_down_proj = self.L__self___mlp_down_proj(mul_2); mul_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:300, code: hidden_states = residual + hidden_states | |
add_2 = add + l__self___mlp_down_proj; add = l__self___mlp_down_proj = None | |
return (add_2,) | |
[2023-07-06 04:20:58,554] torch._dynamo.output_graph.__graph: [DEBUG] TRACED GRAPH | |
__compiled_fn_22 <eval_with_key>.219 opcode name target args kwargs | |
------------- ----------------------------------------- -------------------------------------------------------- ------------------------------------------------ ----------------- | |
placeholder l_stack0_0_ L_stack0_0_ () {} | |
placeholder l_residual_ L_residual_ () {} | |
call_function add <built-in function add> (l_residual_, l_stack0_0_) {} | |
call_method to to (add, torch.float32) {} | |
call_method pow_1 pow (to, 2) {} | |
call_method mean mean (pow_1, -1) {'keepdim': True} | |
call_function add_1 <built-in function add> (mean, 1e-06) {} | |
call_function rsqrt <built-in method rsqrt of type object at 0x7fb940469580> (add_1,) {} | |
call_function mul <built-in function mul> (add, rsqrt) {} | |
get_attr l__self___post_attention_layernorm_weight L__self___post_attention_layernorm_weight () {} | |
call_function mul_1 <built-in function mul> (l__self___post_attention_layernorm_weight, mul) {} | |
call_method to_1 to (mul_1, torch.float16) {} | |
call_module l__self___mlp_gate_proj L__self___mlp_gate_proj (to_1,) {} | |
call_function silu <function silu at 0x7fb945bc5990> (l__self___mlp_gate_proj,) {} | |
call_module l__self___mlp_up_proj L__self___mlp_up_proj (to_1,) {} | |
call_function mul_2 <built-in function mul> (silu, l__self___mlp_up_proj) {} | |
call_module l__self___mlp_down_proj L__self___mlp_down_proj (mul_2,) {} | |
call_function add_2 <built-in function add> (add, l__self___mlp_down_proj) {} | |
output output output ((add_2,),) {} | |
[2023-07-06 04:20:58,555] torch._dynamo.output_graph: [INFO] Step 2: calling compiler function inductor | |
[2023-07-06 04:20:58,602] torch._functorch.aot_autograd.__aot_graphs: [INFO] TRACED GRAPH | |
===== Forward graph 26 ===== | |
<eval_with_key>.223 class <lambda>(torch.nn.Module): | |
def forward(self, arg0_1: f16[16], arg1_1: f16[64, 16], arg2_1: f16[64, 16], arg3_1: f16[16, 64], arg4_1: f16[1, 1, 16], arg5_1: f16[1, 1, 16]): | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:294, code: hidden_states = residual + hidden_states | |
add: f16[1, 1, 16] = torch.ops.aten.add.Tensor(arg5_1, arg4_1); arg5_1 = arg4_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85, code: variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) | |
convert_element_type: f32[1, 1, 16] = torch.ops.prims.convert_element_type.default(add, torch.float32) | |
pow_1: f32[1, 1, 16] = torch.ops.aten.pow.Tensor_Scalar(convert_element_type, 2); convert_element_type = None | |
mean: f32[1, 1, 1] = torch.ops.aten.mean.dim(pow_1, [-1], True); pow_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86, code: hidden_states = hidden_states * torch.rsqrt(variance + self.variance_epsilon) | |
add_1: f32[1, 1, 1] = torch.ops.aten.add.Tensor(mean, 1e-06); mean = None | |
rsqrt: f32[1, 1, 1] = torch.ops.aten.rsqrt.default(add_1); add_1 = None | |
mul: f32[1, 1, 16] = torch.ops.aten.mul.Tensor(add, rsqrt); rsqrt = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88, code: return (self.weight * hidden_states).to(input_dtype) | |
mul_1: f32[1, 1, 16] = torch.ops.aten.mul.Tensor(arg0_1, mul); arg0_1 = mul = None | |
convert_element_type_1: f16[1, 1, 16] = torch.ops.prims.convert_element_type.default(mul_1, torch.float16); mul_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:162, code: return self.down_proj(self.act_fn(self.gate_proj(x)) * self.up_proj(x)) | |
permute: f16[16, 64] = torch.ops.aten.permute.default(arg1_1, [1, 0]); arg1_1 = None | |
view: f16[1, 16] = torch.ops.aten.view.default(convert_element_type_1, [1, 16]) | |
mm: f16[1, 64] = torch.ops.aten.mm.default(view, permute); view = permute = None | |
view_1: f16[1, 1, 64] = torch.ops.aten.view.default(mm, [1, 1, 64]); mm = None | |
# File: /home/ubuntu/miniconda3/lib/python3.10/site-packages/transformers/activations.py:150, code: return nn.functional.silu(input) | |
convert_element_type_2: f32[1, 1, 64] = torch.ops.prims.convert_element_type.default(view_1, torch.float32); view_1 = None | |
sigmoid: f32[1, 1, 64] = torch.ops.aten.sigmoid.default(convert_element_type_2) | |
mul_2: f32[1, 1, 64] = torch.ops.aten.mul.Tensor(convert_element_type_2, sigmoid); convert_element_type_2 = sigmoid = None | |
convert_element_type_3: f16[1, 1, 64] = torch.ops.prims.convert_element_type.default(mul_2, torch.float16); mul_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:162, code: return self.down_proj(self.act_fn(self.gate_proj(x)) * self.up_proj(x)) | |
permute_1: f16[16, 64] = torch.ops.aten.permute.default(arg2_1, [1, 0]); arg2_1 = None | |
view_2: f16[1, 16] = torch.ops.aten.view.default(convert_element_type_1, [1, 16]); convert_element_type_1 = None | |
mm_1: f16[1, 64] = torch.ops.aten.mm.default(view_2, permute_1); view_2 = permute_1 = None | |
view_3: f16[1, 1, 64] = torch.ops.aten.view.default(mm_1, [1, 1, 64]); mm_1 = None | |
mul_3: f16[1, 1, 64] = torch.ops.aten.mul.Tensor(convert_element_type_3, view_3); convert_element_type_3 = view_3 = None | |
permute_2: f16[64, 16] = torch.ops.aten.permute.default(arg3_1, [1, 0]); arg3_1 = None | |
view_4: f16[1, 64] = torch.ops.aten.view.default(mul_3, [1, 64]); mul_3 = None | |
mm_2: f16[1, 16] = torch.ops.aten.mm.default(view_4, permute_2); view_4 = permute_2 = None | |
view_5: f16[1, 1, 16] = torch.ops.aten.view.default(mm_2, [1, 1, 16]); mm_2 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:300, code: hidden_states = residual + hidden_states | |
add_2: f16[1, 1, 16] = torch.ops.aten.add.Tensor(add, view_5); add = view_5 = None | |
return (add_2,) | |
[2023-07-06 04:20:58,645] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 13 | |
[2023-07-06 04:20:58,687] torch._inductor.graph: [DEBUG] Force channels last inputs for 0 conv for the current graph with id 13 | |
[2023-07-06 04:20:58,704] torch._inductor.scheduler: [DEBUG] remove_buffer('buf0') | |
[2023-07-06 04:20:58,713] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_ubuntu/oo/cooxrverhcp2snw6sgvi3xmyupj2eewevtsplvvj674cqwmsnljq.py | |
[2023-07-06 04:20:58,713] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 13 | |
[2023-07-06 04:20:58,715] torch._dynamo.output_graph: [INFO] Step 2: done compiler function inductor | |
[2023-07-06 04:20:58,723] torch._dynamo.convert_frame.__guards: [DEBUG] GUARDS: | |
___check_obj_id(L['self'], 140431662486464) | |
hasattr(L['residual'], '_dynamo_dynamic_indices') == False | |
___check_type_id(L['___stack0'], 7638272) | |
len(L['___stack0']) == 3 | |
hasattr(L['___stack0'][0], '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['___stack0'][1], 7649088) | |
___check_obj_id(L['___stack0'][2], 7649088) | |
___check_obj_id(L['output_attentions'], 7698112) | |
not ___is_grad_enabled() | |
not ___are_deterministic_algorithms_enabled() | |
utils_device.CURRENT_DEVICE == None | |
str(G['torch'].float32) == 'torch.float32' | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks.keys()) == set() | |
___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks, 7509984) | |
set(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks.keys()) == set() | |
[2023-07-06 04:20:58,723] torch._inductor.cudagraph_trees: [DEBUG] Running warmup of function 9 | |
[2023-07-06 04:20:58,724] torch._dynamo.convert_frame.__recompiles: [DEBUG] ('Recompiling function forward in /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py', 'set env var TORCHDYNAMO_REPORT_GUARD_FAILURES=1 to debug further') | |
[2023-07-06 04:20:58,724] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo start tracing forward /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:83 | |
[2023-07-06 04:20:58,725] torch._dynamo.variables.builder: [DEBUG] automatic dynamic L['hidden_states'] size(1) 1 != 1000 | |
[2023-07-06 04:20:58,725] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['hidden_states'] (1, 1, 16) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None] | |
[2023-07-06 04:20:58,726] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:84 | |
[2023-07-06 04:20:58,726] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:58,726] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR dtype [TensorVariable()] | |
[2023-07-06 04:20:58,729] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST input_dtype [TorchVariable(torch.float16)] | |
[2023-07-06 04:20:58,729] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85 | |
[2023-07-06 04:20:58,729] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:58,729] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable()] | |
[2023-07-06 04:20:58,729] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:58,729] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR float32 [GetAttrVariable(TensorVariable(), to), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:58,730] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), to), ConstantVariable(dtype)] | |
[2023-07-06 04:20:58,731] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR pow [TensorVariable()] | |
[2023-07-06 04:20:58,731] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 2 [GetAttrVariable(TensorVariable(), pow)] | |
[2023-07-06 04:20:58,731] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), pow), ConstantVariable(int)] | |
[2023-07-06 04:20:58,732] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR mean [TensorVariable()] | |
[2023-07-06 04:20:58,732] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST -1 [GetAttrVariable(TensorVariable(), mean)] | |
[2023-07-06 04:20:58,732] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST True [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int)] | |
[2023-07-06 04:20:58,732] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('keepdim',) [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int), ConstantVariable(bool)] | |
[2023-07-06 04:20:58,732] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 2 [GetAttrVariable(TensorVariable(), mean), ConstantVariable(int), ConstantVariable(bool), ConstantVariable(tuple)] | |
[2023-07-06 04:20:58,733] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST variance [TensorVariable()] | |
[2023-07-06 04:20:58,733] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86 | |
[2023-07-06 04:20:58,733] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [] | |
[2023-07-06 04:20:58,733] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL torch [TensorVariable()] | |
[2023-07-06 04:20:58,734] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR rsqrt [TensorVariable(), TorchVariable(<module 'torch' from '/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/__init__.py'>)] | |
[2023-07-06 04:20:58,734] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST variance [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>)] | |
[2023-07-06 04:20:58,734] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:58,734] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR variance_epsilon [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable(), NNModuleVariable()] | |
[2023-07-06 04:20:58,734] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_ADD None [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable(), ConstantVariable(float)] | |
[2023-07-06 04:20:58,735] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [TensorVariable(), TorchVariable(<built-in method rsqrt of type object at 0x7fb940469580>), TensorVariable()] | |
[2023-07-06 04:20:58,736] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,736] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:58,737] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88 | |
[2023-07-06 04:20:58,737] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,737] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR weight [NNModuleVariable()] | |
[2023-07-06 04:20:58,737] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['self'].weight (16,) [<DimDynamic.STATIC: 2>] [None] | |
[2023-07-06 04:20:58,738] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:58,738] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_MULTIPLY None [TensorVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,738] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR to [TensorVariable()] | |
[2023-07-06 04:20:58,739] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST input_dtype [GetAttrVariable(TensorVariable(), to)] | |
[2023-07-06 04:20:58,739] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [GetAttrVariable(TensorVariable(), to), TorchVariable(torch.float16)] | |
[2023-07-06 04:20:58,740] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [TensorVariable()] | |
[2023-07-06 04:20:58,740] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo done tracing forward (RETURN_VALUE) | |
[2023-07-06 04:20:58,740] torch._dynamo.symbolic_convert: [DEBUG] RETURN_VALUE triggered compile | |
[2023-07-06 04:20:58,740] torch._dynamo.output_graph: [DEBUG] COMPILING GRAPH due to GraphCompileReason(reason='return_value', user_stack=[<FrameSummary file /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py, line 88 in forward>], graph_break=False) | |
[2023-07-06 04:20:58,741] torch._dynamo.output_graph.__graph_code: [DEBUG] TRACED GRAPH | |
===== __compiled_fn_23 ===== | |
<eval_with_key>.232 class GraphModule(torch.nn.Module): | |
def forward(self, L_hidden_states_ : torch.Tensor): | |
l_hidden_states_ = L_hidden_states_ | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85, code: variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) | |
to = l_hidden_states_.to(torch.float32) | |
pow_1 = to.pow(2); to = None | |
mean = pow_1.mean(-1, keepdim = True); pow_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86, code: hidden_states = hidden_states * torch.rsqrt(variance + self.variance_epsilon) | |
add = mean + 1e-06; mean = None | |
rsqrt = torch.rsqrt(add); add = None | |
mul = l_hidden_states_ * rsqrt; l_hidden_states_ = rsqrt = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88, code: return (self.weight * hidden_states).to(input_dtype) | |
l__self___weight = self.L__self___weight | |
mul_1 = l__self___weight * mul; l__self___weight = mul = None | |
to_1 = mul_1.to(torch.float16); mul_1 = None | |
return (to_1,) | |
[2023-07-06 04:20:58,741] torch._dynamo.output_graph.__graph: [DEBUG] TRACED GRAPH | |
__compiled_fn_23 <eval_with_key>.232 opcode name target args kwargs | |
------------- ---------------- -------------------------------------------------------- --------------------------------- ----------------- | |
placeholder l_hidden_states_ L_hidden_states_ () {} | |
call_method to to (l_hidden_states_, torch.float32) {} | |
call_method pow_1 pow (to, 2) {} | |
call_method mean mean (pow_1, -1) {'keepdim': True} | |
call_function add <built-in function add> (mean, 1e-06) {} | |
call_function rsqrt <built-in method rsqrt of type object at 0x7fb940469580> (add,) {} | |
call_function mul <built-in function mul> (l_hidden_states_, rsqrt) {} | |
get_attr l__self___weight L__self___weight () {} | |
call_function mul_1 <built-in function mul> (l__self___weight, mul) {} | |
call_method to_1 to (mul_1, torch.float16) {} | |
output output output ((to_1,),) {} | |
[2023-07-06 04:20:58,742] torch._dynamo.output_graph: [INFO] Step 2: calling compiler function inductor | |
[2023-07-06 04:20:58,760] torch._functorch.aot_autograd.__aot_graphs: [INFO] TRACED GRAPH | |
===== Forward graph 27 ===== | |
<eval_with_key>.236 class <lambda>(torch.nn.Module): | |
def forward(self, arg0_1: f16[16], arg1_1: f16[1, 1, 16]): | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:85, code: variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) | |
convert_element_type: f32[1, 1, 16] = torch.ops.prims.convert_element_type.default(arg1_1, torch.float32) | |
pow_1: f32[1, 1, 16] = torch.ops.aten.pow.Tensor_Scalar(convert_element_type, 2); convert_element_type = None | |
mean: f32[1, 1, 1] = torch.ops.aten.mean.dim(pow_1, [-1], True); pow_1 = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:86, code: hidden_states = hidden_states * torch.rsqrt(variance + self.variance_epsilon) | |
add: f32[1, 1, 1] = torch.ops.aten.add.Tensor(mean, 1e-06); mean = None | |
rsqrt: f32[1, 1, 1] = torch.ops.aten.rsqrt.default(add); add = None | |
mul: f32[1, 1, 16] = torch.ops.aten.mul.Tensor(arg1_1, rsqrt); arg1_1 = rsqrt = None | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:88, code: return (self.weight * hidden_states).to(input_dtype) | |
mul_1: f32[1, 1, 16] = torch.ops.aten.mul.Tensor(arg0_1, mul); arg0_1 = mul = None | |
convert_element_type_1: f16[1, 1, 16] = torch.ops.prims.convert_element_type.default(mul_1, torch.float16); mul_1 = None | |
return (convert_element_type_1,) | |
[2023-07-06 04:20:58,762] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 14 | |
[2023-07-06 04:20:58,778] torch._inductor.graph: [DEBUG] Force channels last inputs for 0 conv for the current graph with id 14 | |
[2023-07-06 04:20:58,788] torch._inductor.scheduler: [DEBUG] remove_buffer('buf0') | |
[2023-07-06 04:20:58,789] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_ubuntu/vz/cvzk75oefn2tjj3ekdj3gg5a7rpdjqti5tysf4vatbmcjiobtp5i.py | |
[2023-07-06 04:20:58,789] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 14 | |
[2023-07-06 04:20:58,792] torch._dynamo.output_graph: [INFO] Step 2: done compiler function inductor | |
[2023-07-06 04:20:58,795] torch._dynamo.convert_frame.__guards: [DEBUG] GUARDS: | |
___check_obj_id(L['self'], 140431662764176) | |
___check_type_id(L['hidden_states'], 77765680) | |
hasattr(L['hidden_states'], '_dynamo_dynamic_indices') == False | |
not ___is_grad_enabled() | |
not ___are_deterministic_algorithms_enabled() | |
utils_device.CURRENT_DEVICE == None | |
str(G['torch'].float32) == 'torch.float32' | |
[2023-07-06 04:20:58,795] torch._inductor.cudagraph_trees: [DEBUG] Running warmup of function 10 | |
[2023-07-06 04:20:58,795] torch._dynamo.convert_frame.__recompiles: [DEBUG] ('Recompiling function __init__ in <string>', 'set env var TORCHDYNAMO_REPORT_GUARD_FAILURES=1 to debug further') | |
[2023-07-06 04:20:58,796] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo start tracing __init__ <string>:2 | |
[2023-07-06 04:20:58,796] torch._dynamo.variables.builder: [DEBUG] automatic dynamic L['last_hidden_state'] size(1) 1 != 1000 | |
[2023-07-06 04:20:58,796] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['last_hidden_state'] (1, 1, 16) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None] | |
[2023-07-06 04:20:58,797] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line <string>:3 | |
[2023-07-06 04:20:58,797] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST last_hidden_state [] | |
[2023-07-06 04:20:58,797] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [TensorVariable()] | |
[2023-07-06 04:20:58,797] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_ATTR last_hidden_state [TensorVariable(), DataClassVariable()] | |
[2023-07-06 04:20:58,798] torch._dynamo.symbolic_convert: [DEBUG] STORE_ATTR triggered compile | |
Traceback (most recent call last): | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/symbolic_convert.py", line 1195, in STORE_ATTR | |
.call_function(self, [obj, ConstantVariable(inst.argval), val], {}) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/builtin.py", line 591, in call_function | |
result = handler(tx, *args, **kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/builtin.py", line 1103, in call_setattr | |
return obj.call_method(tx, "__setattr__", [name_var, val], {}) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/dicts.py", line 417, in call_method | |
return super().call_method(tx, name, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/dicts.py", line 198, in call_method | |
return super().call_method(tx, name, args, kwargs) | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/variables/base.py", line 260, in call_method | |
raise unimplemented(f"call_method {self} {name} {args} {kwargs}") | |
File "/home/ubuntu/miniconda3/lib/python3.10/site-packages/torch/_dynamo/exc.py", line 140, in unimplemented | |
raise Unsupported(msg) | |
torch._dynamo.exc.Unsupported: call_method DataClassVariable() __setitem__ [ConstantVariable(str), TensorVariable()] {} | |
[2023-07-06 04:20:58,798] torch._dynamo.output_graph: [DEBUG] restore_graphstate: removed 0 nodes | |
[2023-07-06 04:20:58,798] torch._dynamo.output_graph: [DEBUG] COMPILING GRAPH due to GraphCompileReason(reason='store_attr', user_stack=[<FrameSummary file <string>, line 3 in __init__>], graph_break=True) | |
[2023-07-06 04:20:58,802] torch._dynamo.convert_frame.__guards: [DEBUG] GUARDS: | |
___check_type_id(L['self'], 110084784) | |
___check_type_id(L['past_key_values'], 7638272) | |
len(L['past_key_values']) == 2 | |
___check_obj_id(L['self'].attentions, 7649088) | |
hasattr(L['last_hidden_state'], '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['self'].hidden_states, 7649088) | |
___check_obj_id(L['self'].past_key_values, 7649088) | |
___check_obj_id(L['self'].last_hidden_state, 7649088) | |
not ___is_grad_enabled() | |
not ___are_deterministic_algorithms_enabled() | |
utils_device.CURRENT_DEVICE == None | |
[2023-07-06 04:20:58,802] torch._dynamo.convert_frame.__recompiles: [DEBUG] ('Recompiling function <resume in forward> in /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py', 'set env var TORCHDYNAMO_REPORT_GUARD_FAILURES=1 to debug further') | |
[2023-07-06 04:20:58,803] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo start tracing <resume in forward> /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:692 | |
[2023-07-06 04:20:58,803] torch._dynamo.variables.builder: [DEBUG] automatic dynamic L['___stack0'].last_hidden_state size(1) 1 != 1000 | |
[2023-07-06 04:20:58,803] torch._dynamo.variables.builder: [DEBUG] wrap_to_fake L['___stack0'].last_hidden_state (1, 1, 16) [<DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>, <DimDynamic.STATIC: 2>] [None, None, None] | |
[2023-07-06 04:20:58,805] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST ___stack0 [] | |
[2023-07-06 04:20:58,805] torch._dynamo.symbolic_convert: [DEBUG] TRACE JUMP_ABSOLUTE 90 [DataClassVariable()] | |
[2023-07-06 04:20:58,805] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST outputs [DataClassVariable()] | |
[2023-07-06 04:20:58,805] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:704 | |
[2023-07-06 04:20:58,805] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST outputs [] | |
[2023-07-06 04:20:58,805] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST 0 [DataClassVariable()] | |
[2023-07-06 04:20:58,805] torch._dynamo.symbolic_convert: [DEBUG] TRACE BINARY_SUBSCR None [DataClassVariable(), ConstantVariable(int)] | |
[2023-07-06 04:20:58,805] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST hidden_states [TensorVariable()] | |
[2023-07-06 04:20:58,805] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:705 | |
[2023-07-06 04:20:58,805] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST self [] | |
[2023-07-06 04:20:58,805] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR lm_head [NNModuleVariable()] | |
[2023-07-06 04:20:58,806] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST hidden_states [NNModuleVariable()] | |
[2023-07-06 04:20:58,806] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION 1 [NNModuleVariable(), TensorVariable()] | |
[2023-07-06 04:20:58,808] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST logits [TensorVariable()] | |
[2023-07-06 04:20:58,808] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:707 | |
[2023-07-06 04:20:58,808] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [] | |
[2023-07-06 04:20:58,809] torch._dynamo.symbolic_convert: [DEBUG] TRACE STORE_FAST loss [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:58,809] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:708 | |
[2023-07-06 04:20:58,809] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST labels [] | |
[2023-07-06 04:20:58,809] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST None [ConstantVariable(NoneType)] | |
[2023-07-06 04:20:58,809] torch._dynamo.symbolic_convert: [DEBUG] TRACE IS_OP 1 [ConstantVariable(NoneType), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:58,809] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_FALSE 222 [ConstantVariable(bool)] | |
[2023-07-06 04:20:58,809] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:720 | |
[2023-07-06 04:20:58,809] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST return_dict [] | |
[2023-07-06 04:20:58,809] torch._dynamo.symbolic_convert: [DEBUG] TRACE POP_JUMP_IF_TRUE 266 [ConstantVariable(bool)] | |
[2023-07-06 04:20:58,809] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:724 | |
[2023-07-06 04:20:58,809] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_GLOBAL CausalLMOutputWithPast [] | |
[2023-07-06 04:20:58,809] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:725 | |
[2023-07-06 04:20:58,809] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST loss [UserDefinedClassVariable()] | |
[2023-07-06 04:20:58,809] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:726 | |
[2023-07-06 04:20:58,809] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST logits [UserDefinedClassVariable(), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:58,809] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:727 | |
[2023-07-06 04:20:58,809] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST outputs [UserDefinedClassVariable(), ConstantVariable(NoneType), TensorVariable()] | |
[2023-07-06 04:20:58,809] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR past_key_values [UserDefinedClassVariable(), ConstantVariable(NoneType), TensorVariable(), DataClassVariable()] | |
[2023-07-06 04:20:58,810] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:728 | |
[2023-07-06 04:20:58,810] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST outputs [UserDefinedClassVariable(), ConstantVariable(NoneType), TensorVariable(), TupleVariable()] | |
[2023-07-06 04:20:58,810] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR hidden_states [UserDefinedClassVariable(), ConstantVariable(NoneType), TensorVariable(), TupleVariable(), DataClassVariable()] | |
[2023-07-06 04:20:58,810] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:729 | |
[2023-07-06 04:20:58,810] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_FAST outputs [UserDefinedClassVariable(), ConstantVariable(NoneType), TensorVariable(), TupleVariable(), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:58,810] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_ATTR attentions [UserDefinedClassVariable(), ConstantVariable(NoneType), TensorVariable(), TupleVariable(), ConstantVariable(NoneType), DataClassVariable()] | |
[2023-07-06 04:20:58,810] torch._dynamo.symbolic_convert: [DEBUG] TRACE starts_line /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:724 | |
[2023-07-06 04:20:58,810] torch._dynamo.symbolic_convert: [DEBUG] TRACE LOAD_CONST ('loss', 'logits', 'past_key_values', 'hidden_states', 'attentions') [UserDefinedClassVariable(), ConstantVariable(NoneType), TensorVariable(), TupleVariable(), ConstantVariable(NoneType), ConstantVariable(NoneType)] | |
[2023-07-06 04:20:58,810] torch._dynamo.symbolic_convert: [DEBUG] TRACE CALL_FUNCTION_KW 5 [UserDefinedClassVariable(), ConstantVariable(NoneType), TensorVariable(), TupleVariable(), ConstantVariable(NoneType), ConstantVariable(NoneType), ConstantVariable(tuple)] | |
[2023-07-06 04:20:58,811] torch._dynamo.symbolic_convert: [DEBUG] TRACE RETURN_VALUE None [DataClassVariable()] | |
[2023-07-06 04:20:58,811] torch._dynamo.symbolic_convert: [INFO] Step 1: torchdynamo done tracing <resume in forward> (RETURN_VALUE) | |
[2023-07-06 04:20:58,811] torch._dynamo.symbolic_convert: [DEBUG] RETURN_VALUE triggered compile | |
[2023-07-06 04:20:58,811] torch._dynamo.output_graph: [DEBUG] COMPILING GRAPH due to GraphCompileReason(reason='return_value', user_stack=[<FrameSummary file /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py, line 724 in <resume in forward>>], graph_break=False) | |
[2023-07-06 04:20:58,812] torch._dynamo.output_graph.__graph_code: [DEBUG] TRACED GRAPH | |
===== __compiled_fn_25 ===== | |
<eval_with_key>.238 class GraphModule(torch.nn.Module): | |
def forward(self, L_stack0_last_hidden_state : torch.Tensor): | |
l_stack0_last_hidden_state = L_stack0_last_hidden_state | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:705, code: logits = self.lm_head(hidden_states) | |
l__self___lm_head = self.L__self___lm_head(l_stack0_last_hidden_state); l_stack0_last_hidden_state = None | |
return (l__self___lm_head,) | |
[2023-07-06 04:20:58,812] torch._dynamo.output_graph.__graph: [DEBUG] TRACED GRAPH | |
__compiled_fn_25 <eval_with_key>.238 opcode name target args kwargs | |
----------- -------------------------- -------------------------- ----------------------------- -------- | |
placeholder l_stack0_last_hidden_state L_stack0_last_hidden_state () {} | |
call_module l__self___lm_head L__self___lm_head (l_stack0_last_hidden_state,) {} | |
output output output ((l__self___lm_head,),) {} | |
[2023-07-06 04:20:58,812] torch._dynamo.output_graph: [INFO] Step 2: calling compiler function inductor | |
[2023-07-06 04:20:58,823] torch._functorch.aot_autograd.__aot_graphs: [INFO] TRACED GRAPH | |
===== Forward graph 28 ===== | |
<eval_with_key>.242 class <lambda>(torch.nn.Module): | |
def forward(self, arg0_1: f16[32000, 16], arg1_1: f16[1, 1, 16]): | |
# File: /home/ubuntu/accelerated-pytorch-transformers-generation/src/trfs_fast/llama.py:705, code: logits = self.lm_head(hidden_states) | |
permute: f16[16, 32000] = torch.ops.aten.permute.default(arg0_1, [1, 0]); arg0_1 = None | |
view: f16[1, 16] = torch.ops.aten.view.default(arg1_1, [1, 16]); arg1_1 = None | |
mm: f16[1, 32000] = torch.ops.aten.mm.default(view, permute); view = permute = None | |
view_1: f16[1, 1, 32000] = torch.ops.aten.view.default(mm, [1, 1, 32000]); mm = None | |
return (view_1,) | |
[2023-07-06 04:20:58,838] torch._inductor.compile_fx: [INFO] Step 3: torchinductor compiling FORWARDS graph 15 | |
[2023-07-06 04:20:58,844] torch._inductor.graph: [DEBUG] Force channels last inputs for 0 conv for the current graph with id 15 | |
[2023-07-06 04:20:58,846] torch._inductor.graph: [DEBUG] Output code written to: /tmp/torchinductor_ubuntu/7y/c7ycehmc443da67tdrrtaosxvw4ksnvhbehw2tj6bp2msphljdvi.py | |
[2023-07-06 04:20:58,846] torch._inductor.compile_fx: [INFO] Step 3: torchinductor done compiling FORWARDS graph 15 | |
[2023-07-06 04:20:58,848] torch._dynamo.output_graph: [INFO] Step 2: done compiler function inductor | |
[2023-07-06 04:20:58,854] torch._dynamo.convert_frame.__guards: [DEBUG] GUARDS: | |
___check_obj_id(L['self'], 140431662761824) | |
___check_obj_id(L['labels'], 7649088) | |
___check_type_id(L['___stack0'], 110084784) | |
___check_obj_id(L['return_dict'], 7698144) | |
___check_obj_id(L['___stack0'].attentions, 7649088) | |
___check_obj_id(L['___stack0'].hidden_states, 7649088) | |
___check_type_id(L['___stack0'].past_key_values, 7638272) | |
len(L['___stack0'].past_key_values) == 2 | |
hasattr(L['___stack0'].last_hidden_state, '_dynamo_dynamic_indices') == False | |
___check_obj_id(L['___stack0'].past_key_values[0], 7649088) | |
___check_obj_id(L['___stack0'].past_key_values[1], 7649088) | |
not ___is_grad_enabled() | |
not ___are_deterministic_algorithms_enabled() | |
utils_device.CURRENT_DEVICE == None | |
[2023-07-06 04:20:58,854] torch._inductor.cudagraph_trees: [DEBUG] Running warmup of function 11 | |
[2023-07-06 04:20:58,855] torch._inductor.cudagraph_trees: [DEBUG] Recording function 6 of graph recording id 0 | |
[2023-07-06 04:20:58,976] torch._inductor.cudagraph_trees: [DEBUG] Recording function 7 of graph recording id 1 | |
[2023-07-06 04:20:59,088] torch._inductor.cudagraph_trees: [DEBUG] Recording function 8 of graph recording id 2 | |
[2023-07-06 04:20:59,200] torch._inductor.cudagraph_trees: [DEBUG] Recording function 9 of graph recording id 3 | |
[2023-07-06 04:20:59,312] torch._inductor.cudagraph_trees: [DEBUG] Recording function 10 of graph recording id 4 | |
[2023-07-06 04:20:59,423] torch._inductor.cudagraph_trees: [DEBUG] Recording function 11 of graph recording id 5 | |
[2023-07-06 04:20:59,537] torch._inductor.cudagraph_trees: [DEBUG] Recording function 0 of graph recording id 6 | |
[2023-07-06 04:20:59,653] torch._inductor.cudagraph_trees: [DEBUG] Recording function 1 of graph recording id 7 | |
[2023-07-06 04:20:59,771] torch._inductor.cudagraph_trees: [DEBUG] Recording function 2 of graph recording id 8 | |
[2023-07-06 04:20:59,886] torch._inductor.cudagraph_trees: [DEBUG] Recording function 3 of graph recording id 9 | |
[2023-07-06 04:20:59,999] torch._inductor.cudagraph_trees: [DEBUG] Recording function 4 of graph recording id 10 | |
[2023-07-06 04:21:00,110] torch._inductor.cudagraph_trees: [DEBUG] Recording function 5 of graph recording id 11 | |
Warmup/compilation time: 22.32 seconds (2 generate calls) | |
Measuring generate: 0%| | 0/1 [00:00<?, ?it/s][A[A[A | |
Measuring generate: 100%|██████████| 1/1 [00:00<00:00, 167.56it/s] | |
100%|██████████| 1/1 [00:22<00:00, 22.36s/it][A[A | |
100%|██████████| 1/1 [00:22<00:00, 22.36s/it] | |
100%|██████████| 1/1 [00:22<00:00, 22.36s/it][A | |
100%|██████████| 1/1 [00:22<00:00, 22.36s/it] | |
100%|██████████| 1/1 [00:22<00:00, 22.36s/it] | |
100%|██████████| 1/1 [00:22<00:00, 22.36s/it] | |
Preallocate Compile Batch size ... Max GPU memory (MB) Results hash Git hash | |
0 True static 1 ... 29.06 efd05c5d cd00bdb | |
[1 rows x 11 columns] | |
Results also appended to ./results/results_llama.csv | |
[2023-07-06 04:21:00,294] torch._dynamo.utils: [INFO] TorchDynamo compilation metrics: | |
Function Runtimes (s) | |
-------------------------------------- -------------- | |
_compile 19.5125 | |
OutputGraph.call_user_compiler 18.1622 | |
create_aot_dispatcher_function 18.7711 | |
compile_fx.<locals>.fw_compiler_base 17.0007 | |
GraphLowering.run 0.5331 | |
GraphLowering.compile_to_module 14.3102 | |
Scheduler.__init__ 6.1754 | |
Scheduler.codegen 0.2503 | |
WrapperCodeGen.generate 0.0188 | |
cudagraphify 0.0031 | |
CachingAutotuner.benchmark_all_configs 0.8992 |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment