作者
Hao Lu, Xuesong Niu, Jiyao Wang, Yin Wang, Qingyong Hu, Jiaqi Tang, Yuting Zhang, Kaishen Yuan, Bin Huang, Zitong Yu, Dengbo He, Shuiguang Deng, Hao Chen, Yingcong Chen, Shiguang Shan
发表日期
2024
研讨会论文
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition
页码范围
322-331
简介
Multimodal large language models (MLLMs) are designed to process and integrate information from multiple sources such as text speech images and videos. Despite its success in language understanding it is critical to evaluate the performance of downstream tasks for better human-centric applications. This paper assesses the application of MLLMs with 5 crucial abilities for affective computing spanning from visual affective tasks and reasoning tasks. The results show that\gpt has high accuracy in facial action unit recognition and micro-expression detection while its general facial expression recognition performance is not accurate. We also highlight the challenges of achieving fine-grained micro-expression recognition and the potential for further study and demonstrate the versatility and potential of\gpt for handling advanced tasks in emotion recognition and related fields by integrating with task-related agents for more complex tasks such as heart rate estimation through signal processing. In conclusion this paper provides valuable insights into the potential applications and challenges of MLLMs in human-centric computing.
引用总数
学术搜索中的文章
H Lu, X Niu, J Wang, Y Wang, Q Hu, J Tang, Y Zhang… - Proceedings of the IEEE/CVF Conference on Computer …, 2024