You're currently viewing an old version of this dataset. To see the current version, click here.

Mono-ViFI: A Unified Framework for Self-supervised Monocular Depth Estimation

Self-supervised monocular depth estimation has gathered no-table interest since it can liberate training from dependency on depth annotations. In monocular video training case, recent methods only conduct view synthesis between existing camera views, leading to insuffi-cient guidance. To tackle this, we try to synthesize more virtual camera views by flow-based video frame interpolation (VFI), termed as tempo-ral augmentation.

Data and Resources

This dataset has no data

Cite this as

Jinfeng Liu, Lingtong Kong, Bo Li, Zerong Wang, Hong Gu, Jinwei Chen (2024). Dataset: Mono-ViFI: A Unified Framework for Self-supervised Monocular Depth Estimation. https://doi.org/10.57702/7yg6k69l

Private DOI This DOI is not yet resolvable.
It is available for use in manuscripts, and will be published when the Dataset is made public.

Additional Info

Field Value
Created December 2, 2024
Last update December 2, 2024
Defined In https://doi.org/10.48550/arXiv.2407.14126
Author Jinfeng Liu
More Authors
Lingtong Kong
Bo Li
Zerong Wang
Hong Gu
Jinwei Chen
Homepage https://github.com/LiuJF1226/Mono-ViFI