-
RECAP: Principled Recaptioning Improves Image Generation
A text-to-image diffusion model trained on a recaptioned dataset to improve image generation quality and semantic alignment. -
Adding conditional control to text-to-image diffusion models
Adding conditional control to text-to-image diffusion models. -
Learning to follow image editing instructions
Learning to follow image editing instructions. -
Text-to-image generation via masked generative transformers
Text-to-image generation via masked generative transformers. -
Emu: Enhancing Image Generation Models Using Photogenic Needles in a Haystack
Training text-to-image models with web scale image-text pairs enables the generation of a wide range of visual concepts from text. However, these pre-trained models often face... -
Prompting4Debugging: Red-Teaming Text-to-Image Diffusion Models by Finding Pr...
Text-to-image diffusion models, e.g. Stable Diffusion (SD), lately have shown remarkable ability in high-quality content generation, and become one of the representatives for... -
Photorealistic text-to-image diffusion models with deep language understanding
The authors present a photorealistic text-to-image diffusion model with deep language understanding. -
DPOK: Reinforcement Learning for Fine-tuning Text-to-Image Diffusion Models
Learning from human feedback has been shown to improve text-to-image models. These techniques first learn a reward function that captures what humans care about in the task and...