-
Notifications
You must be signed in to change notification settings - Fork 2
Description
Hi,
I'm trying to do a sanity test of reconstructing 1 motion of 16 frames from the DFaust_67 dataset. I'll just detail the flow of code before the mismatch error happens.
The code goes to L132 of generative.py (inside encode_local). At this point, the size of "x" for me is, (1,360,16).
16 being the length of the clip, 360 being 24 joints * 15 params.
Then self.local_encoder(x) is called and the tensor goes through LocalEncoder's forward method.
It goes through 4 layers, with the output size of each layer being:
torch.Size([1, 420, 8])
torch.Size([1, 540, 4])
torch.Size([1, 840, 2])
torch.Size([1, 1680, 1])
After the view operation, the last layer outputs a 1x1680 tensor.
This when passed to self.mu() - L82 of prior.py - gives a size mismatch error as follows:
RuntimeError: mat1 and mat2 shapes cannot be multiplied (1x1680 and 13440x1024)
The Linear layer expects input of size 13440, which is 1680 x args.temporal_scale .
However the output of the last layer I get is of 1680.
I don't know how to account for args.temporal_scale here..
Can you please let me know what I'm doing wrong / how I can fix this?
Thank you so much!
Best,
S