-
Long-term Leap Attention, Short-term Periodic Shift for Video Classification
Video transformer naturally incurs a heavier computation burden than a static vision transformer, as the former processes T times longer sequence than the latter under the... -
Kinetics and Something-Something V2 datasets
The dataset used in the paper for few-shot video classification, containing videos from Kinetics and Something-Something V2 datasets. -
YouTube-8M: A Large-Scale Video Classification Benchmark
YouTube-8M is a large-scale video classification benchmark. -
ImageNet and YouTube-8M
The dataset used in this paper is not explicitly described. However, it is mentioned that the authors used datasets such as ImageNet and YouTube-8M. -
Condensed Movies
The dataset used for text-to-video retrieval and video classification tasks. -
Kinetics400
Video classification is a fundamental problem in many video-based tasks. Applications such as autonomous driving technology, controlling drones and robots are driving the demand... -
Kinetics dataset
The Kinetics dataset is a large-scale action recognition dataset. It contains videos of various actions performed by humans, with annotations of the actions performed. -
Something-Something V1
Video classification is a fundamental problem in many video-based tasks. Applications such as autonomous driving technology, controlling drones and robots are driving the demand... -
Kinetics-600
The Kinetics-600 dataset consists of 392k training videos and 30k validation videos in 600 human action categories. -
Structural Vision Transformer
Structural Vision Transformer (StructViT) is a vision transformer network that leverages structural self-attention (StructSA) to capture correlation structures in images and... -
Kinetics-400
Motion has shown to be useful for video understanding, where motion is typically represented by optical flow. However, computing flow from video frames is very time-consuming.... -
Something-Something V1 & V2
The Something-Something V1 & V2 dataset is a large-scale video dataset created by crowdsourcing. It contains about 100k videos over 174 categories, and the number of videos... -
ActivityNet Captions
The ActivityNet Captions is a benchmark dataset proposed for dense video captioning. There are 20K untrimmed videos in total, and each video has several annotated segments with...