Created
May 14, 2022 02:57
-
-
Save jamesr66a/32390c2c247b063e7e61d5e5313a5b3e to your computer and use it in GitHub Desktop.
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
0%| | 0/25 [00:00<?, ?it/s]The following columns in the training set don't have a corresponding argument in `Pipe.forward` and have been ignored: input_ids, decoder_input_ids, labels. If input_ids, decoder_input_ids, labels are not expected by `Pipe.forward`, you can safely ignore this message. | |
Traceback (most recent call last): | |
File "/fsx/users/jamesreed/hf_t5_gcp_megagpu_pippy/train.py", line 302, in <module> | |
run_worker(args.rank, args.world_size, args) | |
File "/fsx/users/jamesreed/hf_t5_gcp_megagpu_pippy/train.py", line 250, in run_worker | |
main(args, pp_ranks_per_dp_group[rank]) | |
File "/fsx/users/jamesreed/hf_t5_gcp_megagpu_pippy/train.py", line 207, in main | |
run_training(args, model, datasets, data_collator, pippy_driver=pipeline_driver) | |
File "/fsx/users/jamesreed/hf_t5_gcp_megagpu_pippy/train.py", line 125, in run_training | |
trainer.train() | |
File "/fsx/users/jamesreed/transformers/src/transformers/trainer.py", line 1328, in train | |
return inner_training_loop( | |
File "/fsx/users/jamesreed/transformers/src/transformers/trainer.py", line 1539, in _inner_training_loop | |
for step, inputs in enumerate(epoch_iterator): | |
File "/fsx/users/jamesreed/pytorch/torch/utils/data/dataloader.py", line 559, in __next__ | |
data = self._next_data() | |
File "/fsx/users/jamesreed/pytorch/torch/utils/data/dataloader.py", line 599, in _next_data | |
data = self._dataset_fetcher.fetch(index) # may raise StopIteration | |
File "/fsx/users/jamesreed/pytorch/torch/utils/data/_utils/fetch.py", line 52, in fetch | |
return self.collate_fn(data) | |
File "/fsx/users/jamesreed/transformers/src/transformers/trainer_utils.py", line 694, in __call__ | |
return self.data_collator(features) | |
File "/fsx/users/jamesreed/transformers/src/transformers/data/data_collator.py", line 586, in __call__ | |
features = self.tokenizer.pad( | |
File "/fsx/users/jamesreed/transformers/src/transformers/tokenization_utils_base.py", line 2795, in pad | |
raise ValueError( | |
ValueError: You should supply an encoding or a list of encodings to this method that includes input_ids, but you provided [] |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment