The error above was from preprocess step of the training set.
Also for testing (unannotated challenge videos) I was curious whether we are expected to provide predictions for all 30fps frames from each video?
(So basically for preprocessing step of the challenge test set we have to extract all frames inside the video?)
Yes, you are absolutely correct! For the test, we are extracting 30FPS frames. But you only need to report tracking results every 6 frames. Please look at tools/eval_datasets/eval_ego4d_lt_tracking.py line 113 (for some reason, I cannot post link). So, we start from the frame number of the visual crop, and track forward and backward every 6 frames.
One reason for this was because of the frame number conversion issue we had (some annotation does not necessarily start from frame #0, but maybe frame #1). Another reason is different trackers might perform differently in different FPSs, so you could explore running on 5FPS video or 30FPS video.
@haotang Question on test set preprocess step. I meet the error “KeyError: exported_clip_uid” inside extract_clip_ids function. I found that the annotation actually doesn’t include the key “exported_clip_uid”. Should I just skip videos that do not have a ‘exported_clip_uid’ annotation? (Or should i use clip_uid instead?)
@aram It is taking longer for us to investigate why these clips were not released to public. At the same time, please ignore these four clips in the training set. Do you experience any missing clips for the test set?