This repository was archived by the owner on Oct 9, 2023. It is now read-only.
How to use different video classifier backbones #1468
Unanswered
BernhardGlueck
asked this question in
Data / pipelines
Replies: 1 comment
-
Hi @BernhardGlueck Thanks for reporting this! This looks like a bug on our side (partially related to #1328) for the slowfast backends. Regarding the x3d backends, my guess is that they expect the image to be a bit bigger. Could you try creating your datamodule like this: from flash.video.classification.input_transform import VideoClassificationInputTransform
datamodule = VideoClassificationData.from_folders(
...,
transform=VideoClassificationInputTransform(image_size=512),
) That would set the image size to be 512 rather than the default of 244. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I am trying to train a video classifier using the following code:
My input data are thousands of video all with a resolution of 398x224 with 25 fps and a length of exactly 2 seconds each.
I tried different clip samples like random or uniform.
This works fine, but now i want to switch to different backbones.... in the example i am using x3d_xs from the tutorial.
However if i change that to any of the other supported backbones i get different errors:
What am i doing wrong ?
Beta Was this translation helpful? Give feedback.
All reactions