On the Road with GPT-4V (ision): Explorations of Utilizing Visual-Language Model as Autonomous Driving Agent

L Wen, X Yang, D Fu, X Wang, P Cai, X Li… - ICLR 2024 Workshop …, 2024 - openreview.net
The development of autonomous driving technology depends on merging perception,
decision, and control systems. Traditional strategies have struggled to understand complex …

On the road with gpt-4v (ision): Early explorations of visual-language model on autonomous driving

L Wen, X Yang, D Fu, X Wang, P Cai, X Li, T Ma… - arXiv preprint arXiv …, 2023 - arxiv.org
The pursuit of autonomous driving technology hinges on the sophisticated integration of
perception, decision-making, and control systems. Traditional approaches, both data-driven …

Vision language models in autonomous driving: A survey and outlook

X Zhou, M Liu, E Yurtsever, BL Zagar… - IEEE Transactions …, 2024 - ieeexplore.ieee.org
The applications of Vision-Language Models (VLMs) in the field of Autonomous Driving (AD)
have attracted widespread attention due to their outstanding performance and the ability to …

GPT-4V Explorations: Mining Autonomous Driving

Z Li - arXiv preprint arXiv:2406.16817, 2024 - arxiv.org
This paper explores the application of the GPT-4V (ision) large visual language model to
autonomous driving in mining environments, where traditional systems often falter in …

Vision language models in autonomous driving and intelligent transportation systems

X Zhou, M Liu, BL Zagar, E Yurtsever… - arXiv preprint arXiv …, 2023 - arxiv.org
The applications of Vision-Language Models (VLMs) in the fields of Autonomous Driving
(AD) and Intelligent Transportation Systems (ITS) have attracted widespread attention due to …

Co-driver: VLM-based Autonomous Driving Assistant with Human-like Behavior and Understanding for Complex Road Scenes

Z Guo, A Lykov, Z Yagudin, M Konenkov… - arXiv preprint arXiv …, 2024 - arxiv.org
Recent research about Large Language Model based autonomous driving solutions shows
a promising picture in planning and control fields. However, heavy computational resources …

Automated evaluation of large vision-language models on self-driving corner cases

Y Li, W Zhang, K Chen, Y Liu, P Li, R Gao… - arXiv preprint arXiv …, 2024 - arxiv.org
Large Vision-Language Models (LVLMs), due to the remarkable visual reasoning ability to
understand images and videos, have received widespread attention in the autonomous …

[HTML][HTML] DriveLLaVA: Human-Level Behavior Decisions via Vision Language Model

R Zhao, Q Yuan, J Li, Y Fan, Y Li… - Sensors (Basel …, 2024 - ncbi.nlm.nih.gov
Human-level driving is the ultimate goal of autonomous driving. As the top-level decision-
making aspect of autonomous driving, behavior decision establishes short-term driving …

Drivevlm: The convergence of autonomous driving and large vision-language models

X Tian, J Gu, B Li, Y Liu, C Hu, Y Wang, K Zhan… - arXiv preprint arXiv …, 2024 - arxiv.org
A primary hurdle of autonomous driving in urban environments is understanding complex
and long-tail scenarios, such as challenging road conditions and delicate human behaviors …

Embodied understanding of driving scenarios

Y Zhou, L Huang, Q Bu, J Zeng, T Li, H Qiu… - arXiv preprint arXiv …, 2024 - arxiv.org
Embodied scene understanding serves as the cornerstone for autonomous agents to
perceive, interpret, and respond to open driving scenarios. Such understanding is typically …