Dear Authors,
Thank you for your excellent work.
I’m working on a project that requires estimating optical flow for each frame. Because the motions in my data are extremely large, your method—which combines tracking and flow—seems like a great fit.
To obtain per-frame flow, my current approach is to iteratively truncate the video from the beginning and process frames 1…t for each target frame t, while keeping the flow from frame 1 to frame 2. However, this appears to make the model lose context prior to the target frame, which leads to poor tracking for parts of the moving object. Interestingly, adding just one frame before the target frame already resolves the issue.
Do you have any suggestions on how I could approach this problem using your project? Any guidance is appreciated : )
Thank you for your time.