This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
================================================================================ | |
Traceback (most recent call last): | |
File "/usr/local/lib/python3.10/dist-packages/fireworks/util/multiprocessing.py", line 70, in _wrap | |
fn(i, *args) | |
File "/usr/local/lib/python3.10/dist-packages/fireworks/serving/image_generation/coordinator.py", line 32, in _worker_fn | |
worker.init(args) | |
File "/usr/local/lib/python3.10/dist-packages/fireworks/serving/image_generation/worker.py", line 121, in init | |
self.model_manager.warmup_models() | |
File "/usr/local/lib/python3.10/dist-packages/fireworks/serving/image_generation/model_manager.py", line 333, in warmup_models | |
pipeline.text_to_image( |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
ERROR: Error responding to query | |
Traceback (most recent call last): | |
File "/usr/local/lib/python3.10/dist-packages/fastapi_poe/base.py", line 174, in handle_query | |
async for event in self.get_response(query): | |
File "/usr/local/lib/python3.10/dist-packages/fireworks/serving/poe_text_completion_server/fw_poe_server_bot.py", line 114, in get_response | |
async for response in ChatCompletion.acreate( | |
File "/usr/local/lib/python3.10/dist-packages/fireworks/client/base_completion.py", line 145, in _acreate_streaming | |
async for event in _parse_sse_async(response, cls.stream_response_class): | |
File "/usr/local/lib/python3.10/dist-packages/fireworks/client/base_completion.py", line 16, in _parse_sse_async | |
async for line in lines: |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
import torch | |
import torch.fx | |
def foo(x): | |
with torch.autograd.profiler.record_function('fooo'): | |
return torch.relu(x) | |
class RecordFunctionTracer(torch.fx.Tracer): | |
def trace(self, root, concrete_args=None): |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
0%| | 0/25 [00:00<?, ?it/s]The following columns in the training set don't have a corresponding argument in `Pipe.forward` and have been ignored: input_ids, decoder_input_ids, labels. If input_ids, decoder_input_ids, labels are not expected by `Pipe.forward`, you can safely ignore this message. | |
Traceback (most recent call last): | |
File "/fsx/users/jamesreed/hf_t5_gcp_megagpu_pippy/train.py", line 302, in <module> | |
run_worker(args.rank, args.world_size, args) | |
File "/fsx/users/jamesreed/hf_t5_gcp_megagpu_pippy/train.py", line 250, in run_worker | |
main(args, pp_ranks_per_dp_group[rank]) | |
File "/fsx/users/jamesreed/hf_t5_gcp_megagpu_pippy/train.py", line 207, in main | |
run_training(args, model, datasets, data_collator, pippy_driver=pipeline_driver) | |
File "/fsx/users/jamesreed/hf_t5_gcp_megagpu_pippy/train.py", line 125, in run_training | |
trainer.train() |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
###### test.py ##### | |
import torch | |
from transformers import ( | |
Seq2SeqTrainer, | |
Seq2SeqTrainingArguments, | |
T5Config, | |
T5ForConditionalGeneration, | |
DataCollatorForSeq2Seq, | |
AutoTokenizer, | |
set_seed, |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
diff --git a/torch/fx/graph_module.py b/torch/fx/graph_module.py | |
index e2f033d72a..7b3a97991d 100644 | |
--- a/torch/fx/graph_module.py | |
+++ b/torch/fx/graph_module.py | |
@@ -222,6 +222,56 @@ def _assign_attr(from_obj: Any, to_module: torch.nn.Module, target: str): | |
else: | |
setattr(to_module, field, from_obj) | |
+class _WrappedCall: | |
+ def __init__(self, cls, cls_call): |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
import argparse, socket, os | |
import torch | |
import torch.fx | |
import torch.distributed.rpc as rpc | |
import torch.multiprocessing as mp | |
def run_main(args): | |
class MyCode(torch.nn.Module): | |
def forward(self, x): |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
diff --git a/src/transformers/utils/fx.py b/src/transformers/utils/fx.py | |
index 0270295dd..acf3deaf4 100644 | |
--- a/src/transformers/utils/fx.py | |
+++ b/src/transformers/utils/fx.py | |
@@ -485,18 +485,18 @@ class HFTracer(Tracer): | |
_reset_tensor_methods(self.original_methods) | |
- # TODO: keep this until necessary. | |
- # This is necessary because concrete args are added as input to the traced module since |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
diff --git a/test/t5_test.py b/test/t5_test.py | |
index e6b36b8..b72108f 100644 | |
--- a/test/t5_test.py | |
+++ b/test/t5_test.py | |
@@ -34,4 +34,13 @@ t5_input = torch.zeros(bs, seq_length, dtype=torch.long).random_(t5.config.vocab | |
decoder_input_ids = torch.zeros(bs, seq_length, dtype=torch.long).random_(t5.config.vocab_size) | |
t5_output = t5(input_ids=t5_input, decoder_attention_mask=None, decoder_input_ids=decoder_input_ids) | |
-t5_pipe_output = t5_pipe(input_ids=t5_input, decoder_attention_mask=None, decoder_input_ids=decoder_input_ids) | |
+ |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
diff --git a/test/t5_test.py b/test/t5_test.py | |
index e6b36b8..bc7edde 100644 | |
--- a/test/t5_test.py | |
+++ b/test/t5_test.py | |
@@ -15,7 +15,7 @@ for i in range(t5.config.num_layers): | |
for i in range(t5.config.num_decoder_layers): | |
annotate_split_points(t5, {f'decoder.block.{i}': PipeSplitWrapper.SplitPoint.BEGINNING}) | |
-input_names = t5.dummy_inputs.keys() | |
+input_names = t5.dummy_inputs.keys() - ['decoder_attention_mask'] |
NewerOlder