Skip to content

Instantly share code, notes, and snippets.

View jamesr66a's full-sized avatar

James Reed jamesr66a

View GitHub Profile
================================================================================
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/fireworks/util/multiprocessing.py", line 70, in _wrap
fn(i, *args)
File "/usr/local/lib/python3.10/dist-packages/fireworks/serving/image_generation/coordinator.py", line 32, in _worker_fn
worker.init(args)
File "/usr/local/lib/python3.10/dist-packages/fireworks/serving/image_generation/worker.py", line 121, in init
self.model_manager.warmup_models()
File "/usr/local/lib/python3.10/dist-packages/fireworks/serving/image_generation/model_manager.py", line 333, in warmup_models
pipeline.text_to_image(
ERROR: Error responding to query
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/fastapi_poe/base.py", line 174, in handle_query
async for event in self.get_response(query):
File "/usr/local/lib/python3.10/dist-packages/fireworks/serving/poe_text_completion_server/fw_poe_server_bot.py", line 114, in get_response
async for response in ChatCompletion.acreate(
File "/usr/local/lib/python3.10/dist-packages/fireworks/client/base_completion.py", line 145, in _acreate_streaming
async for event in _parse_sse_async(response, cls.stream_response_class):
File "/usr/local/lib/python3.10/dist-packages/fireworks/client/base_completion.py", line 16, in _parse_sse_async
async for line in lines:
import torch
import torch.fx
def foo(x):
with torch.autograd.profiler.record_function('fooo'):
return torch.relu(x)
class RecordFunctionTracer(torch.fx.Tracer):
def trace(self, root, concrete_args=None):
0%| | 0/25 [00:00<?, ?it/s]The following columns in the training set don't have a corresponding argument in `Pipe.forward` and have been ignored: input_ids, decoder_input_ids, labels. If input_ids, decoder_input_ids, labels are not expected by `Pipe.forward`, you can safely ignore this message.
Traceback (most recent call last):
File "/fsx/users/jamesreed/hf_t5_gcp_megagpu_pippy/train.py", line 302, in <module>
run_worker(args.rank, args.world_size, args)
File "/fsx/users/jamesreed/hf_t5_gcp_megagpu_pippy/train.py", line 250, in run_worker
main(args, pp_ranks_per_dp_group[rank])
File "/fsx/users/jamesreed/hf_t5_gcp_megagpu_pippy/train.py", line 207, in main
run_training(args, model, datasets, data_collator, pippy_driver=pipeline_driver)
File "/fsx/users/jamesreed/hf_t5_gcp_megagpu_pippy/train.py", line 125, in run_training
trainer.train()
###### test.py #####
import torch
from transformers import (
Seq2SeqTrainer,
Seq2SeqTrainingArguments,
T5Config,
T5ForConditionalGeneration,
DataCollatorForSeq2Seq,
AutoTokenizer,
set_seed,
diff --git a/torch/fx/graph_module.py b/torch/fx/graph_module.py
index e2f033d72a..7b3a97991d 100644
--- a/torch/fx/graph_module.py
+++ b/torch/fx/graph_module.py
@@ -222,6 +222,56 @@ def _assign_attr(from_obj: Any, to_module: torch.nn.Module, target: str):
else:
setattr(to_module, field, from_obj)
+class _WrappedCall:
+ def __init__(self, cls, cls_call):
import argparse, socket, os
import torch
import torch.fx
import torch.distributed.rpc as rpc
import torch.multiprocessing as mp
def run_main(args):
class MyCode(torch.nn.Module):
def forward(self, x):
diff --git a/src/transformers/utils/fx.py b/src/transformers/utils/fx.py
index 0270295dd..acf3deaf4 100644
--- a/src/transformers/utils/fx.py
+++ b/src/transformers/utils/fx.py
@@ -485,18 +485,18 @@ class HFTracer(Tracer):
_reset_tensor_methods(self.original_methods)
- # TODO: keep this until necessary.
- # This is necessary because concrete args are added as input to the traced module since
diff --git a/test/t5_test.py b/test/t5_test.py
index e6b36b8..b72108f 100644
--- a/test/t5_test.py
+++ b/test/t5_test.py
@@ -34,4 +34,13 @@ t5_input = torch.zeros(bs, seq_length, dtype=torch.long).random_(t5.config.vocab
decoder_input_ids = torch.zeros(bs, seq_length, dtype=torch.long).random_(t5.config.vocab_size)
t5_output = t5(input_ids=t5_input, decoder_attention_mask=None, decoder_input_ids=decoder_input_ids)
-t5_pipe_output = t5_pipe(input_ids=t5_input, decoder_attention_mask=None, decoder_input_ids=decoder_input_ids)
+
diff --git a/test/t5_test.py b/test/t5_test.py
index e6b36b8..bc7edde 100644
--- a/test/t5_test.py
+++ b/test/t5_test.py
@@ -15,7 +15,7 @@ for i in range(t5.config.num_layers):
for i in range(t5.config.num_decoder_layers):
annotate_split_points(t5, {f'decoder.block.{i}': PipeSplitWrapper.SplitPoint.BEGINNING})
-input_names = t5.dummy_inputs.keys()
+input_names = t5.dummy_inputs.keys() - ['decoder_attention_mask']