Skip to content

Instantly share code, notes, and snippets.

@jamesr66a
Created May 14, 2022 02:57
Show Gist options
  • Star 0 You must be signed in to star a gist
  • Fork 0 You must be signed in to fork a gist
  • Save jamesr66a/32390c2c247b063e7e61d5e5313a5b3e to your computer and use it in GitHub Desktop.
Save jamesr66a/32390c2c247b063e7e61d5e5313a5b3e to your computer and use it in GitHub Desktop.
0%| | 0/25 [00:00<?, ?it/s]The following columns in the training set don't have a corresponding argument in `Pipe.forward` and have been ignored: input_ids, decoder_input_ids, labels. If input_ids, decoder_input_ids, labels are not expected by `Pipe.forward`, you can safely ignore this message.
Traceback (most recent call last):
File "/fsx/users/jamesreed/hf_t5_gcp_megagpu_pippy/train.py", line 302, in <module>
run_worker(args.rank, args.world_size, args)
File "/fsx/users/jamesreed/hf_t5_gcp_megagpu_pippy/train.py", line 250, in run_worker
main(args, pp_ranks_per_dp_group[rank])
File "/fsx/users/jamesreed/hf_t5_gcp_megagpu_pippy/train.py", line 207, in main
run_training(args, model, datasets, data_collator, pippy_driver=pipeline_driver)
File "/fsx/users/jamesreed/hf_t5_gcp_megagpu_pippy/train.py", line 125, in run_training
trainer.train()
File "/fsx/users/jamesreed/transformers/src/transformers/trainer.py", line 1328, in train
return inner_training_loop(
File "/fsx/users/jamesreed/transformers/src/transformers/trainer.py", line 1539, in _inner_training_loop
for step, inputs in enumerate(epoch_iterator):
File "/fsx/users/jamesreed/pytorch/torch/utils/data/dataloader.py", line 559, in __next__
data = self._next_data()
File "/fsx/users/jamesreed/pytorch/torch/utils/data/dataloader.py", line 599, in _next_data
data = self._dataset_fetcher.fetch(index) # may raise StopIteration
File "/fsx/users/jamesreed/pytorch/torch/utils/data/_utils/fetch.py", line 52, in fetch
return self.collate_fn(data)
File "/fsx/users/jamesreed/transformers/src/transformers/trainer_utils.py", line 694, in __call__
return self.data_collator(features)
File "/fsx/users/jamesreed/transformers/src/transformers/data/data_collator.py", line 586, in __call__
features = self.tokenizer.pad(
File "/fsx/users/jamesreed/transformers/src/transformers/tokenization_utils_base.py", line 2795, in pad
raise ValueError(
ValueError: You should supply an encoding or a list of encodings to this method that includes input_ids, but you provided []
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment