-
Flickr30K and MSCOCO
The dataset used in the paper is Flickr30K and MSCOCO, which are used for image-text matching and image captioning tasks. -
Stacked Cross Attention
The dataset used in the paper is Stacked Cross Attention for Image-Text Matching. -
Language Models with Image Descriptors
The Language Models with Image Descriptors dataset, which is used for evaluating the performance of the InstructVid2Vid model. -
SpatialSense
A dataset for visual spatial relationship classification (VSRC) with nine well-defined spatial relations. -
Compositional Visual Genome
The Compositional Visual Genome (ComVG) dataset is a reconstructed dataset of the Visual Genome (Krishna et al., 2017) dataset, containing 108,007 images annotated with 2.3... -
Visual Genome
The Visual Genome dataset is a large-scale visual question answering dataset, containing 1.5 million images, each with 15-30 annotated entities, attributes, and relationships.