10 datasets found

Tags: diffusion models

Filter Results
  • InstructVid2Vid dataset

    The dataset used for training the InstructVid2Vid model, which consists of video-instruction-edited video triplets.
  • MotionFollower

    The dataset used in the paper is not explicitly described, but it is mentioned that the authors collect 3K videos (60-90 seconds long) from the internet to train their model.
  • Emu Video Edit Training Dataset

    The Emu Video Edit model's training dataset, containing 1600 videos with 7 editing instructions each.
  • Emu Video Edit Dataset

    The dataset used for training the Emu Video Edit model, containing 1600 videos.
  • ControlVideo

    ControlVideo is a general framework to utilize T2I diffusion models for one-shot video editing, which incorporates additional conditions such as edge maps, the key frame and...
  • Tune-A-Video

    The dataset used in the paper for video editing tasks
  • EI2 model for text-driven video editing

    The dataset used in the paper is not explicitly described, but it is mentioned that the authors used the DAVIS dataset and the Pexels website to gather face videos.
  • Show-1

    The Show-1 dataset is a text-to-video diffusion model.
  • Zeroscope

    The dataset used in this paper for text-to-video generation, consisting of short video clips.
  • Davis and WebVid datasets

    The dataset used in the paper is not explicitly described, but it is mentioned that the authors used 26 text-video pairs from the public DAVIS and WebVid datasets.