Mono-ViFI: A Unified Framework for Self-supervised Monocular Depth Estimation

Self-supervised monocular depth estimation has gathered no-table interest since it can liberate training from dependency on depth annotations. In monocular video training case, recent methods only conduct view synthesis between existing camera views, leading to insuffi-cient guidance. To tackle this, we try to synthesize more virtual camera views by flow-based video frame interpolation (VFI), termed as tempo-ral augmentation.

Data and Resources

Cite this as

Jinfeng Liu, Lingtong Kong, Bo Li, Zerong Wang, Hong Gu, Jinwei Chen (2024). Dataset: Mono-ViFI: A Unified Framework for Self-supervised Monocular Depth Estimation. https://doi.org/10.57702/7yg6k69l

DOI retrieved: December 2, 2024

Additional Info

Field Value
Created December 2, 2024
Last update December 2, 2024
Defined In https://doi.org/10.48550/arXiv.2407.14126
Author Jinfeng Liu
More Authors
Lingtong Kong
Bo Li
Zerong Wang
Hong Gu
Jinwei Chen
Homepage https://github.com/LiuJF1226/Mono-ViFI