Hi, thank you for the implementation of the C3D net.
I'm currently trying to train the model on the ucf dataset. As they described in the paper, I choose several (10/12/14/16) frames from a clip, unsqueeze each on dimension 1, then concatenate all of them. So the input tensor has a torch.Size([3, 10/12/14/16, 224, 224]).
Then I got a error:
''
File "/home/cxing95/anaconda3/lib/python3.6/site-packages/torch/nn/functional.py", line 371, in max_pool3d
ret = torch._C._nn.max_pool3d(input, kernel_size, stride, padding, dilation, ceil_mode)
RuntimeError: Given input size: (512x1x14x14). Calculated output size: (512x0x8x8). Output size is too small at /opt/conda/conda-bld/pytorch_1524584710464/work/aten/src/THCUNN/generic/VolumetricDilatedMaxPooling.cu:105
''
I'm not sure if it raised because of the input size. But could you tell what's the input should be like to train or use this net. That will help a lot.
Thank you.