-
Notifications
You must be signed in to change notification settings - Fork 12
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Can you please point to the code that tracks during inference? #18
Comments
Our MeMOTR is an RNN-like model, processing the video frame-by-frame, like word-by-word in RNNs. So, in theory, the processing length is unlimited. Therefore, videos longer than sample length frames (during training) will not cause any difference in inference, still frame-by-frame. So, we do not connect several shorter tracks into an overall trajectory. For the time step However, inconsistent lengths during training and inference can indeed cause issues for the model. I further discuss this topic in my recent work. |
Hello, may I ask what you said about the inconsistent length during training and inference. Can you explain in depth? My training dataset is 300 frames and the test set is 18,000 frames. When my test set reaches 1000 frames, there will be serious tracking confusion. Is this caused by the inconsistent length of the training dataset and the test set? Have you encountered such problems in previous experiments? |
I think that's not what I mean by inconsistent length. Let me explain it: In your description, although your dataset is 300 frames long, for training, it's not different from a 5-frame clip. So, the inconsistency is not between 300 and 18000. |
I am confused about how the tracking is performed during inference for videos longer than sample length frames? What part of the code connects those shorter tracks?
The text was updated successfully, but these errors were encountered: