-
CRD-CGAN: Category-Consistent and Relativistic Constraints for Diverse Text-t...
Generating photo-realistic images from a text description is a challenging problem in computer vision. Previ- ous works have shown promising performance to generate synthetic... -
Concept Sliders
The dataset used for concept sliders, including Age, Surprise, and Hair sliders. -
Text-to-Image Diffusion Models
The dataset used for text-to-image diffusion models, including Bluefire, Paintings, 3D, and Origami styles. -
HPSv2 dataset
The HPSv2 dataset is a text-image pair dataset containing 3200 prompts and their corresponding images. -
A-SDM: Accelerating Stable Diffusion through Model Assembly and Feature Inher...
The Stable Diffusion Model (SDM) is a preva- lent and effective model for text-to-image (T2I) and image-to-image (I2I) generation. -
Custom Diffusion
The dataset used in this paper is a large-scale text-to-image diffusion model, which consists of 35 subjects with unique pets and objects. -
Continuous 3D Words
A dataset of images with fine-grained control over several 3D-aware attributes, including time-of-day illumination, bird wing orientation, dollyzoom effect, and object poses. -
ABC-6K dataset
The ABC-6K dataset includes prompts with at least two color words modifying different objects. -
DVMP dataset
The DVMP dataset features a diverse set of objects and diverse modifiers including colors, textures, etc. -
AnE dataset
The AnE dataset comprises three benchmarks: Animal-Animal, Animal-Object, and Object-Object. -
Parti-Prompts
The dataset used in the paper for testing the TextCraftor model. -
OpenPrompt1
The dataset used in the paper for training and testing the TextCraftor model. -
TextCraftor: Your Text Encoder Can be Image Quality Controller
TextCraftor is a stable and powerful framework to fine-tune the pre-trained text encoder to improve the text-to-image generation. -
GreenStableYolo
The dataset used in the paper is not explicitly described, but it is mentioned that the authors used Stable Diffusion and Yolo to optimize the parameters and prompts for... -
Stable Diffusion-1.5
The dataset used in the paper is not explicitly described, but it is mentioned that the authors used a 20-step DDIM sampler and a 13-step DDIM sampler. -
Sber-MoVQGAN
A large dataset of text-image pairs collected online. -
MUSE: Text-to-Image Generation via Masked Generative Transformers
MUSE is a text-to-image generation model that uses masked generative transformers. -
Photorealistic text-to-image diffusion models with deep language understanding
The authors present a photorealistic text-to-image diffusion model with deep language understanding.