8 datasets found

Groups: Vision-and-Language Navigation

Filter Results
  • Iterative Vision-and-Language Navigation in Continuous Environments

    Iterative Vision-and-Language Navigation in Continuous Environments (IR2R-CE) benchmark for evaluating language-guided agents navigating in a persistent environment over time.
  • Iterative Room-to-Room

    Iterative Room-to-Room (IR2R) benchmark for evaluating language-guided agents navigating in a persistent environment over time.
  • Iterative Vision-and-Language Navigation

    Iterative Vision-and-Language Navigation (IVLN) paradigm for evaluating language-guided agents navigating in a persistent environment over time.
  • Room-to-Room

    The Room-to-Room dataset is a photo-realistic dataset for vision-and-language navigation, where agents navigate through indoor environments based on natural language instructions.
  • Room-to-Room (R2R) dataset

    The Room-to-Room (R2R) dataset is a benchmark for vision-and-language navigation tasks. It consists of 7,189 paths sampled from its navigation graphs, each with three...
  • R2R

    The dataset used in the paper for vision-and-language navigation tasks.
  • REVERIE dataset

    The REVERIE dataset is a dataset of household tasks in an indoor environment. It contains images annotated with natural language instructions including the referring expressions...
  • Vision-and-Language Navigation

    The Vision-and-Language Navigation (VLN) task gives a global natural sentence I = {w0,..., wl} as an instruction, where wi is a word token while the l is the length of the...