[PDF][PDF] Measuring the Visual Hallucination in ChatGPT on Visually Deceptive Images

L Ping, Y Gu, L Feng - files.osf.io
The evaluation of visual hallucinations in multimodal AI models is novel and significant
because it addresses a critical gap in understanding how AI systems interpret deceptive …

RITUAL: Random Image Transformations as a Universal Anti-hallucination Lever in LVLMs

S Woo, J Jang, D Kim, Y Choi, C Kim - arXiv preprint arXiv:2405.17821, 2024 - arxiv.org
Recent advancements in Large Vision Language Models (LVLMs) have revolutionized how
machines understand and generate textual responses based on visual inputs. Despite their …

Mitigating hallucination in large multi-modal models via robust instruction tuning

F Liu, K Lin, L Li, J Wang, Y Yacoob… - The Twelfth International …, 2023 - openreview.net
Despite the promising progress in multi-modal tasks, current large multi-modal models
(LMMs) are prone to hallucinating inconsistent descriptions with respect to the associated …

Assessing visual hallucinations in vision-enabled large language models

P Lu, L Huang, T Wen, T Shi - 2024 - researchsquare.com
Recent advancements in vision-enabled large language models have prompted a renewed
interest in evaluating their capabilities and limitations when interpreting complex visual data …

AIGCs Confuse AI Too: Investigating and Explaining Synthetic Image-induced Hallucinations in Large Vision-Language Models

Y Gao, J Wang, Z Lin, J Sang - arXiv preprint arXiv:2403.08542, 2024 - arxiv.org
The evolution of Artificial Intelligence Generated Contents (AIGCs) is advancing towards
higher quality. The growing interactions with AIGCs present a new challenge to the data …

Holistic analysis of hallucination in gpt-4v (ision): Bias and interference challenges

C Cui, Y Zhou, X Yang, S Wu, L Zhang, J Zou… - arXiv preprint arXiv …, 2023 - arxiv.org
While GPT-4V (ision) impressively models both visual and textual information
simultaneously, it's hallucination behavior has not been systematically assessed. To bridge …

Pensieve: Retrospect-then-compare mitigates visual hallucination

D Yang, B Cao, G Chen, C Jiang - arXiv preprint arXiv:2403.14401, 2024 - arxiv.org
Multi-modal Large Language Models (MLLMs) demonstrate remarkable success across
various vision-language tasks. However, they suffer from visual hallucination, where the …

Visualizing Invariant Features in Vision Models

F Sammani, B Joukovsky… - 2023 24th International …, 2023 - ieeexplore.ieee.org
Explainable AI is important for improving transparency, accountability, trust, and ethical
considerations in AI systems, and for enabling users to make informed decisions based on …

Post hoc visual interpretation using a deep learning-based smooth feature network

I Naseem Abbasi, TM Madni, MK Sohail, UI Janjua… - Soft Computing, 2023 - Springer
Interpreting deep learning (DL) models is difficult due to the complexity of their internal
representations. Given the inherent lack of interpretability, it is challenging to identify the …

Boosting Cross-task Transferability of Adversarial Patches with Visual Relations

T Ma, S Li, Y Xiao, S Liu - arXiv preprint arXiv:2304.05402, 2023 - arxiv.org
The transferability of adversarial examples is a crucial aspect of evaluating the robustness of
deep learning systems, particularly in black-box scenarios. Although several methods have …