-
Pre-training tasks for embedding-based large-scale retrieval
This paper proposes a pre-training task for embedding-based large-scale retrieval. -
Playing Lottery Tickets with Vision and Language
Large-scale pre-training has recently revolutionized vision-and-language (VL) research. Models such as LXMERT and UNITER have achieved state-of-the-art performance across a wide... -
Chinese CLIP
A vision-language pre-training dataset, Chinese CLIP, which consists of 100 million image-text pairs. -
RAMM: Retrieval-augmented Biomedical Visual Question Answering
A retrieval-augmented pretrain-and-finetune paradigm for biomedical VQA which includes a high-quality image-text pairs PMCPM, a pre-trained multi-modal model, and a novel... -
DataComp-1B
The dataset used in the paper is also DataComp-1B, which is a large-scale dataset for training next-generation image-text models. -
LAION-400M and LAION-5B
The dataset used in the paper is LAION-400M and LAION-5B, which are large-scale datasets for training next-generation image-text models. -
HuBERT Framework
The dataset used in this paper is a self-supervised audio pre-training framework called HuBERT. -
BERT: Pre-training of deep bidirectional transformers for language understanding
This paper proposes BERT, a pre-trained deep bidirectional transformer for language understanding. -
Various Datasets
The datasets used in the paper are described as follows: WikiMIA, BookMIA, Temporal Wiki, Temporal arXiv, ArXiv-1 month, Multi-Webdata, LAION-MI, Gutenberg. -
OSCAR Dataset
The dataset used in the paper is a large corpus of real-world programs for pre-training a neural network model to learn better code representation. -
Unified language model pre-training for natural language understanding and ge...
A unified language model pre-training for natural language understanding and generation.