M2Chat: Empowering VLM for Multimodal LLM Interleaved

M2Chat is a novel unified multimodal LLM framework for generating interleaved text-image conversation across various scenarios.

Data and Resources

Cite this as

Xiaowei Chi, Rongyu Zhang, Zhengkai Jiang, Yijiang Liu, Yatian Wang, Xingqun Qi, Wenhan Luo, Peng Gao, Shanghang Zhang, Qifeng Liu (2024). Dataset: M2Chat: Empowering VLM for Multimodal LLM Interleaved. https://doi.org/10.57702/u22e5hc4

DOI retrieved: December 2, 2024

Additional Info

Field Value
Created December 2, 2024
Last update December 2, 2024
Author Xiaowei Chi
More Authors
Rongyu Zhang
Zhengkai Jiang
Yijiang Liu
Yatian Wang
Xingqun Qi
Wenhan Luo
Peng Gao
Shanghang Zhang
Qifeng Liu
Homepage https://mattie-e.github.io/M2Chat.github.io