Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Downsampling process is different from that described in the paper #22

Open
light1726 opened this issue Aug 22, 2019 · 3 comments
Open

Comments

@light1726
Copy link

Thanks for sharing the code and you did a great job.
I noticed that in the paper the downsampling process on the temporal axis is different for the forward sequence and the backward sequence. But it seems that the downsampling operation for the forward sequence in the code follows exactly the process described in the paper for the backward sequence. I'm quite confused because these two processes (what described in the code and in the paper) seems to behave differently for that they encode different contextual information.
Since the code is more up-to-date, does the downsampling process in the code is better?
image

codes.append(torch.cat((out_forward[:,i+self.freq-1,:],out_backward[:,i,:]), dim=-1))

@light1726
Copy link
Author

By the way, do you cut the input sequence into fixed-length which is the multiple of the downsampling frequency during training? if so how long is the fixed-length?

@auspicious3000
Copy link
Owner

auspicious3000 commented Aug 23, 2019

Thanks.
The code is correct.
2 seconds.

@light1726
Copy link
Author

I see, thanks for the answer.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants