Holistic Autonomous Driving Understanding by Bird's-Eye-View Injected Multi-Modal Large Models

X Ding, J Han, H Xu, X Liang… - Proceedings of the …, 2024 - openaccess.thecvf.com
The rise of multimodal large language models (MLLMs) has spurred interest in language-
based driving tasks. However existing research typically focuses on limited tasks and often …

Hilm-d: Towards high-resolution understanding in multimodal large language models for autonomous driving

X Ding, J Han, H Xu, W Zhang, X Li - arXiv preprint arXiv:2309.05186, 2023 - arxiv.org
Autonomous driving systems generally employ separate models for different tasks resulting
in intricate designs. For the first time, we leverage singular multimodal large language …

A survey on multimodal large language models for autonomous driving

C Cui, Y Ma, X Cao, W Ye, Y Zhou… - Proceedings of the …, 2024 - openaccess.thecvf.com
With the emergence of Large Language Models (LLMs) and Vision Foundation Models
(VFMs), multimodal AI systems benefiting from large models have the potential to equally …

Lmdrive: Closed-loop end-to-end driving with large language models

H Shao, Y Hu, L Wang, G Song… - Proceedings of the …, 2024 - openaccess.thecvf.com
Despite significant recent progress in the field of autonomous driving modern methods still
struggle and can incur serious accidents when encountering long-tail unforeseen events …

Drivegpt4: Interpretable end-to-end autonomous driving via large language model

Z Xu, Y Zhang, E Xie, Z Zhao, Y Guo, KKY Wong… - arXiv preprint arXiv …, 2023 - arxiv.org
In the past decade, autonomous driving has experienced rapid development in both
academia and industry. However, its limited interpretability remains a significant unsolved …

Vlaad: Vision and language assistant for autonomous driving

SY Park, MJ Lee, JH Kang, H Choi… - Proceedings of the …, 2024 - openaccess.thecvf.com
While interpretable decision-making is pivotal in autonomous driving, research integrating
natural language models remains a relatively untapped. To address this, we introduce a …

Drivemlm: Aligning multi-modal large language models with behavioral planning states for autonomous driving

W Wang, J Xie, CY Hu, H Zou, J Fan, W Tong… - arXiv preprint arXiv …, 2023 - arxiv.org
Large language models (LLMs) have opened up new possibilities for intelligent agents,
endowing them with human-like thinking and cognitive abilities. In this work, we delve into …

Driving with llms: Fusing object-level vector modality for explainable autonomous driving

L Chen, O Sinavski, J Hünermann, A Karnsund… - arXiv preprint arXiv …, 2023 - arxiv.org
Large Language Models (LLMs) have shown promise in the autonomous driving sector,
particularly in generalization and interpretability. We introduce a unique object-level …

LimSim++: A Closed-Loop Platform for Deploying Multimodal LLMs in Autonomous Driving

D Fu, W Lei, L Wen, P Cai, S Mao, M Dou, B Shi… - arXiv preprint arXiv …, 2024 - arxiv.org
The emergence of Multimodal Large Language Models ((M) LLMs) has ushered in new
avenues in artificial intelligence, particularly for autonomous driving by offering enhanced …

Vision language models in autonomous driving: A survey and outlook

X Zhou, M Liu, E Yurtsever, BL Zagar… - IEEE Transactions …, 2024 - ieeexplore.ieee.org
The applications of Vision-Language Models (VLMs) in the field of Autonomous Driving (AD)
have attracted widespread attention due to their outstanding performance and the ability to …