Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Input dimension limitation #14

Open
tawsinDOTuddin opened this issue Apr 28, 2022 · 5 comments
Open

Input dimension limitation #14

tawsinDOTuddin opened this issue Apr 28, 2022 · 5 comments

Comments

@tawsinDOTuddin
Copy link

The code seems to be only work on a specific input resolution which the model trained on. If I want to test it with dimension other than 432x240, I can but I have to train the model with that specific dimension. right?

@ruiliu-ai
Copy link
Owner

Correct.
If you want to test your model with arbitrary-size input, you can remove the positional embedding layer. That's the key reason why the input dimension must be fixed.

@tawsinDOTuddin
Copy link
Author

I think it is same issue with your another work "Decoupled Spatial-Temporal Transformer for Video Inpainting"....what i should do if i test that model with arbitrary-size input?

@ruiliu-ai
Copy link
Owner

Remove positional embedding, and edit this line to be 1/4 * (h, w) where (h, w) can be arbitrary input/output image size
https://github.com/ruiliu-ai/FuseFormer/blob/master/model/fuseformer.py#L119

@ruiliu-ai
Copy link
Owner

Please refer to ruiliu-ai/DSTT#4

@tawsinDOTuddin
Copy link
Author

Thank you for your response :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants