-
Notifications
You must be signed in to change notification settings - Fork 43
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
About Pixel Shuffle #21
Comments
Hi @chautuankien , thanks for you interest in our work.
PixelShuffle downscales the spatial resolution ( For instance, if you apply a 3x3(xC) kernel to a
From what I've understood, I think you're talking about optical flow based models that use the input images for warping. Our model focuses on direct synthesis without flow-based warping, so the method is very different. There are pros and cons for each method, but flow-based works are more popular these days, to be frank.. |
Thank you so much for your reply. So, for the first question, how PS works is like Pooling layer, right? For example, in case of Max Pooling of stride 2, it chooses the maximum value in a 2x2 grid, to down-sampling For the second question, from what I've understood, is your method a CNN-based method? You will use CNNs to directly synthesize the intermediate frame. Another question is, why did you choose to down-shuffle only once but not more? (just like an encoder-decoder-based network, where Pooling layer is used more than once to down-sampling the data) |
It is very interesting that you use Pixel Shuffle and Channel Attention for motion estimation without estimating optical flow.
I want to ask that in the paper you said that using Pixel Shuffle to maintain the large receptive field, so I want to ask how PS can do that.
One more question, in VFI, I usually see that people will use again the input images to reconstruct the color for the middle. So how just by applying Up Shuffle you can synthesize the middle frame?
Thank you.
The text was updated successfully, but these errors were encountered: