The features have been updated as of 2022-06-07. Please re-download them if you have used them before this date. There was a bug in the feature generation pipeline which had caused the model to not be fed the video input correctly. You can find the paths to the old features by reading the description below.
Pre-extracted feature vectors are available for every video in the
dataset. They can be accessed with the EGO4D
CLI. Please consult the table below for the appropriate
Want to Add a Model?
Refer to the features README on the Ego4D github.
If you need support in running the job to extract features, please open an issue on the github repository.
Here is a table of the features pre-extracted from Ego4D. These features are extracted from the canonical videos. Canonical videos are all 30FPS.
Window Size and Stride are in frames.
|Feature Type||Dataset(s) Trained On||Model Arch||Window Size||Stride||Model Weights Location||Notes|
|Kinetics 400||SlowFast 8x8 (R101 backbone)||32||16||torchub path: facebookresearch/pytorchvideo/slowfast_r101|
|Kinetics 400 / ImageNet-1K||Omnivore (swin L); video head||32||16||https://github.com/facebookresearch/omnivore#model-zoo|
|Kinetics 400 / ImageNet-1K||Omnivore (swin L); image head||1||5||https://github.com/facebookresearch/omnivore#model-zoo|
|Kinetics 400 / ImageNet-1K||Omnivore (swin L); video head||32||16||https://github.com/facebookresearch/omnivore#model-zoo||FP16 variant of |
There is additionally
omnivore_video_deprecated for the features released before 2022-06-07
Features are extracted in a moving window fashion. At every extraction
point the model sees the next Window Size (
W) frames (i.e. at frame
i the model sees features
[i, i + W) frames). The window starts at
frame 0, and then is offset by the stride until the end of the video
There is a boundary condition where the last window may extend past
the video. In this case, the extraction point is backed up such that a
W frames from the video is used. This occurs when the
number of frames in the canonical video is not divisible by the stride.
Example Window Stride
Let's say a video has 39 frames. The frames for extraction will be (in frame numbers):
- [0, 31]
- [7, 38] which is “back-padded” from [16, 47] to fit the last window
- Read the README on the Ego4D github to run or use the code yourself
- UniformClipSampler from PytorchVideo is used