Saliency prediction in the deep learning era: Successes and limitations

A Borji - IEEE transactions on pattern analysis and machine …, 2019 - ieeexplore.ieee.org
Visual saliency models have enjoyed a big leap in performance in recent years, thanks to
advances in deep learning and large scale annotated data. Despite enormous effort and …

Electromagnetic nanocommunication networks: Principles, applications, and challenges

MH Kabir, SMR Islam, AP Shrestha, F Ali… - IEEE …, 2021 - ieeexplore.ieee.org
Nanoscale devices, also called nanomachines, form communication networks and
cooperate with each other so that they can be used to perform complicated tasks. Such …

Yourefit: Embodied reference understanding with language and gesture

Y Chen, Q Li, D Kong, YL Kei, SC Zhu… - Proceedings of the …, 2021 - openaccess.thecvf.com
We study the machine's understanding of embodied reference: One agent uses both
language and gesture to refer to an object to another agent in a shared physical …

Tidying deep saliency prediction architectures

N Reddy, S Jain, P Yarlagadda… - 2020 IEEE/RSJ …, 2020 - ieeexplore.ieee.org
Learning computational models for visual attention (saliency estimation) is an effort to inch
machines/robots closer to human visual cognitive abilities. Data-driven efforts have …

Eqa-mx: Embodied question answering using multimodal expression

MM Islam, A Gladstone, R Islam… - The Twelfth International …, 2023 - openreview.net
Humans predominantly use verbal utterances and nonverbal gestures (eg, eye gaze and
pointing gestures) in their natural interactions. For instance, pointing gestures and verbal …

Interpreting multimodal referring expressions in real time

D Whitney, M Eldon, J Oberlin… - 2016 IEEE International …, 2016 - ieeexplore.ieee.org
Humans communicate about objects using language, gesture, and context, fusing
information from multiple modalities over time. Robots need to interpret this communication …

CAESAR: An embodied simulator for generating multimodal referring expression datasets

MM Islam, R Mirzaiee, A Gladstone… - Advances in Neural …, 2022 - proceedings.neurips.cc
Humans naturally use verbal utterances and nonverbal gestures to refer to various objects
(known as $\textit {referring expressions} $) in different interactional scenarios. As collecting …

Joint attention by gaze interpolation and saliency

Z Yücel, AA Salah, Ç Meriçli, T Meriçli… - IEEE Transactions …, 2013 - ieeexplore.ieee.org
Joint attention, which is the ability of coordination of a common point of reference with the
communicating party, emerges as a key factor in various interaction scenarios. This paper …

Complementary effects of gaze direction and early saliency in guiding fixations during free viewing

A Borji, D Parks, L Itti - Journal of vision, 2014 - jov.arvojournals.org
Gaze direction provides an important and ubiquitous communication channel in daily
behavior and social interaction of humans and some animals. While several studies have …

Attentional mechanisms for socially interactive robots–a survey

JF Ferreira, J Dias - IEEE Transactions on Autonomous Mental …, 2014 - ieeexplore.ieee.org
This review intends to provide an overview of the state of the art in the modeling and
implementation of automatic attentional mechanisms for socially interactive robots. Humans …