Hello.
I have received my aws cli license from ego4d yesterday, and I’m trying to download the “EgoTracks” dataset.
I can successsfully download the viz and annotations but egotracks videos is failing.
ego4d --output_directory="~/scratch/data/tracking/ego4d" --datasets egotracks output: botocore.exceptions.ClientError: An error occurred (403) when calling the HeadObject operation: Forbidden
Also I wanted to ask how large (in GB) the EgoTracks dataset is? can only find it consists of 5.9K videos but I couldn’t find anywhere it states the actual size.
I am running into the same problem. If I do not use the region in config file then it throws ValueError: Invalid endpoint and if I use us-east-1 or us-east-2 then it throws the error stated above. Did you find any solution?
Hello I have not been able to resolve this. Also i think --datasets annotations_540ss also doesn’t work… Please let me know where i can find the downscaled 540ss annotations!
I think the --datasets egotracks only gives the annotation json files.
So do I have to download the video data by the following command??? ego4d --output_directory ./ --datasets full_scale --benchmark EM --version v2
Hi, which annotation_path are you using (train, val or test)?
This should not happen with the challenge test set, but if it does, please let us know!
For the train and val, we are working on pushing an updated preprocess that should solve the problem. The workaround is: You can simple ignore these clips (should be less than 1%). We don’t have the exported_clip_uid for these frames because of conversion error.
Thank you so much for the fast reply! I think I will try to re-download only the clip dataset in the meanwhile. Please let me know when the preprocess script is cleaned :). Thanks @haotang!
@haotang
Just as an FYI. Skipping 113 videos...Total 3433 to be processed ... Is the number of videos that don’t have “exported_clip_uid” field in annotations
I have tried running the preprocess code but it hangs. Also I can no longer cd or ls into the drive storing the data. Currently I am using a 4TB ssd to store all the data. Could you please provide reference on what is the total disk space that is required to run the preprocess script?
Thanks. I have a 4 TB ssd, and I have a problem where the extraction code hangs in the middle, and I cannot ls into my ssd. (probably the extraction process/thread is not exitting??)
So I have cancelled - restarted multiple times but know the extracted frames folder is ~ 3.8 TB. (disk space is basically full)
Do you suspect anything going wrong?
Major problems
frame extraction process hanging (probably due to 2? but not sure)
Disk space requirement > 8 TB
I did some calculation where
each video ~ 8 min with 30fps = 8 * 60 * 30 = 14400 frames. (I checked and the extracted folder actually has 14400 frames)
Each frame ~ 200 KB
Each video image folder (extracted frames) = 14400 * 200 KB = 2.88 GB.
Train set includes 3000 videos which leads to 8.6 TB disk space for extracted frames.
Yes, true. We annotate at 5FPS, so it should be fine if only extracting those frames. If you extract at 30 FPS, the disk space is not enough. Please take a look at the pull request here: Egotracks fix by tanghaotommy · Pull Request #42 · EGO4D/episodic-memory · GitHub. EgoTracks/tools/preprocess/extract_ego4d_clip_annotated_frames.py only extracts annotated frames.