Hello,
first of all thank u for this nice video DataLoader.
I want to implement this into my project but currently I encounter into problems.
Like u described I preprocessed my video into frames and created the .txt file.
If u want now to load into my project I get a RuntimeError : RuntimeError: Expected 3D (unbatched) or 4D (batched) input to conv2d, but got input of size: [125, 5, 3, 224, 224]
The 125 is my batch size, 5 my num_segments, and so on.
I tried to reshape my video batch like this:
batch_size, frames, channels, height, width = video.shapevideo = video.reshape(batch_size * frames, channels, height, width)
But then I get problems with my labels in the batch:
ValueError: Target size (torch.Size([64, 1])) must be the same as input size (torch.Size([320, 1]))
Hello, first of all thank u for this nice video DataLoader.
I want to implement this into my project but currently I encounter into problems. Like u described I preprocessed my video into frames and created the .txt file. If u want now to load into my project I get a RuntimeError :
RuntimeError: Expected 3D (unbatched) or 4D (batched) input to conv2d, but got input of size: [125, 5, 3, 224, 224]
The 125 is my batch size, 5 my num_segments, and so on.I tried to reshape my video batch like this:
batch_size, frames, channels, height, width = video.shape
video = video.reshape(batch_size * frames, channels, height, width)
But then I get problems with my labels in the batch:ValueError: Target size (torch.Size([64, 1])) must be the same as input size (torch.Size([320, 1]))
Do u know how to fix it or did I do sth wrong?
here's my code- part I am currently using:
Thanks in advance :)