-
Conceptual Captions
The dataset used in the paper "Scaling Laws of Synthetic Images for Model Training". The dataset is used for supervised image classification and zero-shot classification tasks. -
Measuring Machine Intelligence through Visual Question Answering
Measuring machine intelligence through visual question answering. -
VQA: Visual Question Answering
Visual Question Answering (VQA) has emerged as a prominent multi-discipline research problem in both academia and industry. -
Hierarchical Question-Image Co-Attention for Visual Question Answering
A number of recent works have proposed attention models for Visual Question Answering (VQA) that generate spatial maps highlighting image regions relevant to answering the... -
SBU Captions
The SBU Captions dataset is a large-scale image-text dataset used for vision-language pre-training. -
Amazon Berkeley Objects Dataset (ABO)
The Amazon Berkeley Objects Dataset (ABO) is a public available e-commerce dataset with multiple images per product. -
Visual Genome
The Visual Genome dataset is a large-scale visual question answering dataset, containing 1.5 million images, each with 15-30 annotated entities, attributes, and relationships. -
Microsoft COCO
The Microsoft COCO dataset was used for training and evaluating the CNNs because it has become a standard benchmark for testing algorithms aimed at scene understanding and...