generated from rom1504/python-template
-
Notifications
You must be signed in to change notification settings - Fork 67
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Example on how to handle a padded video frame in downstream network? #315
Comments
Also, I'm still getting errors with the current code even after passing Error logs:
My code:
|
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Hello,
I'm using WebVid10M dataset, and passing the following decoder args:
Since I'm passing 16 frames, I often get errors saying
ValueError: video clip not long enough for decoding
since there maybe some shorter videos in the dataset.Thus, if I pass in
pad_frames = True
in abovedecoder_kwargs
, I believe shorter clips would be padded up.Could you give some example code/snippet showing what would be the optimal way to utilize this padded batch of data when processing with our neural nets? Like do we get access to the start of padding, or the pad token? Do we need to use some kind of masking in our forward pass to ignore the padded tokens or just leave the padded data as it is? I was planning on using UNet3D network from diffusers to process the data.
The text was updated successfully, but these errors were encountered: