Last active
March 30, 2020 17:22
-
-
Save chaonan99/c3dd34fe93402371eeab74a44678d558 to your computer and use it in GitHub Desktop.
A pytorch multiprocessing problem
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
import os | |
import torch | |
class Dataset(torch.utils.data.Dataset): | |
arg = {'batch_size': 1} | |
def __init__(self, arg): | |
print('__init__') | |
self.arg.update(arg) | |
# self.arg = self.arg | |
print(self.arg) | |
def _worker_init_fn(self, *args): | |
print('worker init') | |
print(self.arg) | |
def get_dataloader(self): | |
return torch.utils.data.DataLoader(self, batch_size=None, | |
num_workers=3, | |
worker_init_fn=self._worker_init_fn, | |
pin_memory=True, | |
multiprocessing_context='spawn') | |
def __getitem__(self, idx): | |
return 0 | |
def __len__(self): | |
return 5 | |
def main(): | |
dataloader = Dataset({'batch_size': 2}).get_dataloader() | |
for _ in dataloader: | |
pass | |
if __name__ == '__main__': | |
main() |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Strange enough! If I add line 11 it works...