Z Tang, Z Yang, M Khademi, Y Liu, C Zhu… - arXiv preprint arXiv …, 2023 - arxiv.org
We present CoDi-2, a versatile and interactive Multimodal Large Language Model (MLLM)
that can follow complex multimodal interleaved instructions, conduct in-context learning …