Closed silexxx closed 4 years ago
From src/GANcing/../pix2pixHD/util/visualizer.py:26: The name tf.summary.FileWriter is deprecated. Please use tf.compat.v1.summary.FileWriter instead.
actually this issue I am encountering with all the GitHub everybody dance now repositories I tried to fix it but still getting stuck can u help me with this, please Thank you
This looks more like a pytorch/cuda problem than a bug in the code to me...
src/config/train_opt.py
to 1 (or maybe 0)There is nothing I can change in the code that would solve this unfortunately
yes, I would look more into it whats causing the error to thank you so much.
yes I got the problem solved by changing the tensorflow-gpu from 1.14.0 to 1.15.2 in colab. Thank you again. for the quick response.
nceNow-master/AMLD2020-Dirty-GANcing$ python src/GANcing/train_pose2vid.py -t data/targets/target -r danesh CustomDatasetDataLoader dataset [AlignedDataset] was created
training images = 203
GlobalGenerator( (model): Sequential( (0): ReflectionPad2d((3, 3, 3, 3)) (1): Conv2d(18, 64, kernel_size=(7, 7), stride=(1, 1)) (2): InstanceNorm2d(64, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) (3): ReLU(inplace=True) (4): Conv2d(64, 128, kernel_size=(3, 3), stride=(2, 2), padding=(1, 1)) (5): InstanceNorm2d(128, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) (6): ReLU(inplace=True) (7): Conv2d(128, 256, kernel_size=(3, 3), stride=(2, 2), padding=(1, 1)) (8): InstanceNorm2d(256, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) (9): ReLU(inplace=True) (10): Conv2d(256, 512, kernel_size=(3, 3), stride=(2, 2), padding=(1, 1)) (11): InstanceNorm2d(512, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) (12): ReLU(inplace=True) (13): Conv2d(512, 1024, kernel_size=(3, 3), stride=(2, 2), padding=(1, 1)) (14): InstanceNorm2d(1024, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) (15): ReLU(inplace=True) (16): ResnetBlock( (conv_block): Sequential( (0): ReflectionPad2d((1, 1, 1, 1)) (1): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(1, 1)) (2): InstanceNorm2d(1024, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) (3): ReLU(inplace=True) (4): ReflectionPad2d((1, 1, 1, 1)) (5): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(1, 1)) (6): InstanceNorm2d(1024, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) ) ) (17): ResnetBlock( (conv_block): Sequential( (0): ReflectionPad2d((1, 1, 1, 1)) (1): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(1, 1)) (2): InstanceNorm2d(1024, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) (3): ReLU(inplace=True) (4): ReflectionPad2d((1, 1, 1, 1)) (5): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(1, 1)) (6): InstanceNorm2d(1024, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) ) ) (18): ResnetBlock( (conv_block): Sequential( (0): ReflectionPad2d((1, 1, 1, 1)) (1): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(1, 1)) (2): InstanceNorm2d(1024, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) (3): ReLU(inplace=True) (4): ReflectionPad2d((1, 1, 1, 1)) (5): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(1, 1)) (6): InstanceNorm2d(1024, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) ) ) (19): ResnetBlock( (conv_block): Sequential( (0): ReflectionPad2d((1, 1, 1, 1)) (1): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(1, 1)) (2): InstanceNorm2d(1024, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) (3): ReLU(inplace=True) (4): ReflectionPad2d((1, 1, 1, 1)) (5): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(1, 1)) (6): InstanceNorm2d(1024, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) ) ) (20): ResnetBlock( (conv_block): Sequential( (0): ReflectionPad2d((1, 1, 1, 1)) (1): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(1, 1)) (2): InstanceNorm2d(1024, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) (3): ReLU(inplace=True) (4): ReflectionPad2d((1, 1, 1, 1)) (5): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(1, 1)) (6): InstanceNorm2d(1024, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) ) ) (21): ResnetBlock( (conv_block): Sequential( (0): ReflectionPad2d((1, 1, 1, 1)) (1): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(1, 1)) (2): InstanceNorm2d(1024, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) (3): ReLU(inplace=True) (4): ReflectionPad2d((1, 1, 1, 1)) (5): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(1, 1)) (6): InstanceNorm2d(1024, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) ) ) (22): ResnetBlock( (conv_block): Sequential( (0): ReflectionPad2d((1, 1, 1, 1)) (1): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(1, 1)) (2): InstanceNorm2d(1024, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) (3): ReLU(inplace=True) (4): ReflectionPad2d((1, 1, 1, 1)) (5): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(1, 1)) (6): InstanceNorm2d(1024, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) ) ) (23): ResnetBlock( (conv_block): Sequential( (0): ReflectionPad2d((1, 1, 1, 1)) (1): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(1, 1)) (2): InstanceNorm2d(1024, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) (3): ReLU(inplace=True) (4): ReflectionPad2d((1, 1, 1, 1)) (5): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(1, 1)) (6): InstanceNorm2d(1024, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) ) ) (24): ResnetBlock( (conv_block): Sequential( (0): ReflectionPad2d((1, 1, 1, 1)) (1): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(1, 1)) (2): InstanceNorm2d(1024, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) (3): ReLU(inplace=True) (4): ReflectionPad2d((1, 1, 1, 1)) (5): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(1, 1)) (6): InstanceNorm2d(1024, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) ) ) (25): ConvTranspose2d(1024, 512, kernel_size=(3, 3), stride=(2, 2), padding=(1, 1), output_padding=(1, 1)) (26): InstanceNorm2d(512, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) (27): ReLU(inplace=True) (28): ConvTranspose2d(512, 256, kernel_size=(3, 3), stride=(2, 2), padding=(1, 1), output_padding=(1, 1)) (29): InstanceNorm2d(256, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) (30): ReLU(inplace=True) (31): ConvTranspose2d(256, 128, kernel_size=(3, 3), stride=(2, 2), padding=(1, 1), output_padding=(1, 1)) (32): InstanceNorm2d(128, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) (33): ReLU(inplace=True) (34): ConvTranspose2d(128, 64, kernel_size=(3, 3), stride=(2, 2), padding=(1, 1), output_padding=(1, 1)) (35): InstanceNorm2d(64, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) (36): ReLU(inplace=True) (37): ReflectionPad2d((3, 3, 3, 3)) (38): Conv2d(64, 3, kernel_size=(7, 7), stride=(1, 1)) (39): Tanh() ) ) MultiscaleDiscriminator( (scale0_layer0): Sequential( (0): Conv2d(21, 64, kernel_size=(4, 4), stride=(2, 2), padding=(2, 2)) (1): LeakyReLU(negative_slope=0.2, inplace=True) ) (scale0_layer1): Sequential( (0): Conv2d(64, 128, kernel_size=(4, 4), stride=(2, 2), padding=(2, 2)) (1): InstanceNorm2d(128, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) (2): LeakyReLU(negative_slope=0.2, inplace=True) ) (scale0_layer2): Sequential( (0): Conv2d(128, 256, kernel_size=(4, 4), stride=(2, 2), padding=(2, 2)) (1): InstanceNorm2d(256, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) (2): LeakyReLU(negative_slope=0.2, inplace=True) ) (scale0_layer3): Sequential( (0): Conv2d(256, 512, kernel_size=(4, 4), stride=(1, 1), padding=(2, 2)) (1): InstanceNorm2d(512, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) (2): LeakyReLU(negative_slope=0.2, inplace=True) ) (scale0_layer4): Sequential( (0): Conv2d(512, 1, kernel_size=(4, 4), stride=(1, 1), padding=(2, 2)) ) (scale1_layer0): Sequential( (0): Conv2d(21, 64, kernel_size=(4, 4), stride=(2, 2), padding=(2, 2)) (1): LeakyReLU(negative_slope=0.2, inplace=True) ) (scale1_layer1): Sequential( (0): Conv2d(64, 128, kernel_size=(4, 4), stride=(2, 2), padding=(2, 2)) (1): InstanceNorm2d(128, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) (2): LeakyReLU(negative_slope=0.2, inplace=True) ) (scale1_layer2): Sequential( (0): Conv2d(128, 256, kernel_size=(4, 4), stride=(2, 2), padding=(2, 2)) (1): InstanceNorm2d(256, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) (2): LeakyReLU(negative_slope=0.2, inplace=True) ) (scale1_layer3): Sequential( (0): Conv2d(256, 512, kernel_size=(4, 4), stride=(1, 1), padding=(2, 2)) (1): InstanceNorm2d(512, eps=1e-05, momentum=0.1, affine=False, track_running_stats=False) (2): LeakyReLU(negative_slope=0.2, inplace=True) ) (scale1_layer4): Sequential( (0): Conv2d(512, 1, kernel_size=(4, 4), stride=(1, 1), padding=(2, 2)) ) (downsample): AvgPool2d(kernel_size=3, stride=2, padding=[1, 1]) ) WARNING:tensorflow:From src/GANcing/../pix2pixHD/util/visualizer.py:26: The name tf.summary.FileWriter is deprecated. Please use tf.compat.v1.summary.FileWriter instead.
create web directory ./checkpoints/danesh/web... torch_shm_managertorch_shm_manager: error while loading shared libraries: libnvToolsExt.so.1: cannot open shared object file: No such file or directory : error while loading shared libraries: libnvToolsExt.so.1: cannot open shared object file: No such file or directory Traceback (most recent call last): File "src/GANcing/train_pose2vid.py", line 183, in
train_pose2vid(args.target_dir, args.run_name, args.temporal_smoothing)
File "src/GANcing/train_pose2vid.py", line 61, in train_pose2vid
for i, data in enumerate(dataset, start=epoch_iter):
File "/home/daneshwar/Desktop/pytorch-EverybodyDanceNow-master/AMLD2020-Dirty-GANcing/env/lib/python3.7/site-packages/torch/utils/data/dataloader.py", line 819, in next
return self._process_data(data)
File "/home/daneshwar/Desktop/pytorch-EverybodyDanceNow-master/AMLD2020-Dirty-GANcing/env/lib/python3.7/site-packages/torch/utils/data/dataloader.py", line 846, in _process_data
data.reraise()
File "/home/daneshwar/Desktop/pytorch-EverybodyDanceNow-master/AMLD2020-Dirty-GANcing/env/lib/python3.7/site-packages/torch/_utils.py", line 385, in reraise
raise self.exc_type(msg)
RuntimeError: Caught RuntimeError in DataLoader worker process 0.
Original Traceback (most recent call last):
File "/home/daneshwar/Desktop/pytorch-EverybodyDanceNow-master/AMLD2020-Dirty-GANcing/env/lib/python3.7/site-packages/torch/utils/data/_utils/worker.py", line 178, in _worker_loop
data = fetcher.fetch(index)
File "/home/daneshwar/Desktop/pytorch-EverybodyDanceNow-master/AMLD2020-Dirty-GANcing/env/lib/python3.7/site-packages/torch/utils/data/_utils/fetch.py", line 47, in fetch
return self.collate_fn(data)
File "/home/daneshwar/Desktop/pytorch-EverybodyDanceNow-master/AMLD2020-Dirty-GANcing/env/lib/python3.7/site-packages/torch/utils/data/_utils/collate.py", line 74, in default_collate
return {key: default_collate([d[key] for d in batch]) for key in elem}
File "/home/daneshwar/Desktop/pytorch-EverybodyDanceNow-master/AMLD2020-Dirty-GANcing/env/lib/python3.7/site-packages/torch/utils/data/_utils/collate.py", line 74, in
return {key: default_collate([d[key] for d in batch]) for key in elem}
File "/home/daneshwar/Desktop/pytorch-EverybodyDanceNow-master/AMLD2020-Dirty-GANcing/env/lib/python3.7/site-packages/torch/utils/data/_utils/collate.py", line 53, in default_collate
storage = elem.storage()._new_shared(numel)
File "/home/daneshwar/Desktop/pytorch-EverybodyDanceNow-master/AMLD2020-Dirty-GANcing/env/lib/python3.7/site-packages/torch/storage.py", line 128, in _new_shared
return cls._new_using_filename(size)
RuntimeError: error executing torch_shm_manager at "/home/daneshwar/Desktop/pytorch-EverybodyDanceNow-master/AMLD2020-Dirty-GANcing/env/lib/python3.7/site-packages/torch/bin/torch_shm_manager" at /pytorch/torch/lib/libshm/core.cpp:99