Workshop on Nonverbal Cues for Human-Robot Cooperative Intelligence

About the Workshop

This workshop is dedicated to discussing computational methods for sensing and recognition of nonverbal cues and internal states in the wild to realize cooperative intelligence between humans and intelligent systems. We gather researchers from different expertise, yet having the common goal, motivation, and resolve to explore and tackle this delicate issue considering the practicality of industrial applications. We are calling for papers to discuss novel methods to realize human-robot cooperative intelligence by sensing and understanding humans’ behavior, internal states, and to generate empathetic interactions.

  • Human internal state inference, e.g., cognitive, emotional, intention models.
  • Recognition of nonverbal cues, e.g., gaze and attention, body language, para-language.
  • Multi-modal sensing fusion for scene perception.
  • Nonverbal behavior generation for robots/agents, e.g., gaze salience, gesture.
  • Synchronization of nonverbal and verbal behavior
  • Learning algorithms for cooperative intelligence, e.g., imitation learning.
  • Generative and adversarial algorithms to enhance human-robot interaction, e.g., LLMs, diffusion models, VLMs.
  • Empathetic interaction between humans and intelligent systems.
  • Robust sensing of facial and body key points.
  • Social interaction dynamics modeling, e.g., harmony level, engagements.
  • Personalization of intelligent systems from nonverbal cues and trust evaluation.
  • Applications of cooperative intelligence in the wild.

Keywords: "Human: Face, gaze, body, pose, gesture, movement, attention, cognitivestate, emotion state, intention, empathy, Environment: Object"

Secondary subject: "Human-Robot cooperative intelligence", "Nonverbal cues recognition from audiovisual", "Human internal state inference from multi-modality", "Vision applications and systems", "Human-Object interaction and scene understanding"

Sponsors

Organizers

News updates

Jul 1st Workshop web page was launched.
Jul 2nd Submission can be made on EasyChair.
Aug 14th Workshop paper submission deadline is extended to Aug 24th (final) due to multiple requests.

Call for Papers

Submission Guidelines

We invite authors to submit unpublished papers (2-4 pages excluding references) to our workshop, to be presented at a workshop session upon acceptance. Submissions will undergo a peer-review process by the workshop's program committee and accepted papers will be invited to present their works at the workshop (see presentation format).

We are pleased to announce that award will be given to the best paper accepted by this workshop.

Important Dates

  • August 14, 2024 PST

    Paper submission deadline

  • August 24, 2024 PST

    Extended paper submission deadline (final)

  • September 14, 2024 PST

    Notification of acceptance

  • September 28, 2024 PST

    Camera-ready paper submission deadline

  • September 28, 2024 PST

    Pre-recorded video submission deadline

  • October 14, 2024

    Workshop day

Submission Instructions

Please use the IEEE conferences paper format to write your manuscript. Please submit your paper electronically through the workshop's EasyChair submission system.

Presentation Format

Accepted papers should be presented in three-way presentation approach to foster active participation
  • Spotlight talks (6 mins talk, Q&A in the poster session)
  • In-person A0 posters for in-depth discussions
  • Short pre-recorded videos (about 2 minutes) to be uploaded on the workshop webpage

Publication Format

Authors are recommended to archive their papers and inform workshop organizers once this procedure is completed.
Accepted papers which have been archived will be hosted on the workshop webpage.

Program

We plan a half-day event for 4 hours, including talks by two invited speakers, and one interactive session. For participants who could not attend in person, we will disseminate the papers and pre-recorded videos on our workshop page, which also consists of a comment section for Q&A.

  • 09:00 09:02 Welcome and opening remarks (2 mins)
  • 09:02 09:42 Invited talk I (40 mins, including 5 mins Q&A)
  • 09:42 10:30 Spotlight talks for accepted workshop and invited IROS 2024 papers (6 + 2 papers, 6 mins each)
  • 10:30 11:00 Coffee break and poster session (30 mins)
  • 11:00 11:40 Invited talk II (40 mins, including 5 mins Q&A)
  • 11:40 12:20 Invited talk III (40 mins, including 5 mins Q&A)
  • 12:20 12:55 Interactive session (35 mins)
  • 12:55 13:00 Awards and closing remarks (5 mins)

Invited Speakers

We intend to have speakers from different ethnic backgrounds, countries, and career stages. Specifically, we confirmed the attendance of one speaker from the industry, and the other speaker from the academia is pending confirmation.

Satoshi Shigemi, Honda Research Institute Japan, Japan.
Satoshi Shigemi is the President of Honda Research Institute Japan. Since 1987, he has been conducting research on robots and control systems at Honda R&D Co. In 2000, he was the Senior Chief Engineer and project lead for the research and development of ASIMO, the humanoid robot. He then developed a high-altitude survey robot for the Fukushima Daiichi Nuclear Power Plant. He has published many papers about human-robot interaction. Satoshi has tentatively confirmed to be the speaker at our workshop. The preliminary title is “Co-Existence with Intelligent Machines”.
Yukie Nagai, The University of Tokyo, Japan.
Yukie Nagai is a Project Professor at the International Research Center for Neurointelligence at the University of Tokyo. She earned her Ph.D. in Engineering from Osaka University in 2004, after which she worked at the National Institute of Information and Communications Technology, Bielefeld University, and then Osaka University. Since 2019, she has been leading the Cognitive Developmental Robotics Lab at the University of Tokyo. Her research encompasses cognitive developmental robotics, computational neuroscience, and assistive technologies for developmental disorders. Dr. Nagai employs computational methods to investigate the underlying neural mechanisms involved in social cognitive development. In acknowledgment of her work, she received the titles of "World's 50 Most Renowned Women in Robotics" in 2020 and "35 Women in Robotics Engineering and Science" in 2022, among other recognitions.The tentative title: Emergence of Cooperative Intelligence through Embodied Predictive Processing
Angelica Lim, Simon Fraser University
Dr. Angelica Lim is the Director of the Rosie Lab, and an Assistant Professor in the School of Computing Science at Simon Fraser University (SFU). Previously, she led the Emotion and Expressivity teams for the Pepper humanoid robot at SoftBank Robotics. She received her B.Sc. in Computing Science with Artificial Intelligence Specialization from SFU and a Ph.D. and M.Sc. in Computer Science (Intelligence Science) from Kyoto University, Japan. She and her team have received Best Paper in Entertainment Robotics and Cognitive Robotics Awards at IROS 2011 and 2022, and Best Demo and LBR at HRI 2021 and 2023. She has been featured on the BBC, TEDx, hosted a TV documentary on robotics, and was recently featured in Forbes 20 Leading Women in AI. Her research interests include multimodal machine learning, affective computing, and human-robot interaction. The title of her talk is "Multimodal Social Signal Processing for Human-Robot Interaction"

Motivation and Background

Humans can perceive social cues and the interaction context of another human to infer the internal states including cognitive and emotional states, empathy, and intention. This unique ability to infer internal states leads to effective social interaction between humans desirable in many intelligent systems such as collaborative and social robots, and humanmachine interaction systems. However, it is challenging for machines to perceive human states under noisy real-world settings, which are usually measured by noninvasive sensors. Recent works investigating the potential solutions for the estimation of human states under controlled conditions using facial features with the off-the-shelf camera by leveraging deep learning methods. This workshop aims to bring interdisciplinary researchers across computer vision, artificial intelligence, robotics, and human-computer interaction together to share current research achievements and discuss future research directions for human behavior and state understanding, and their potential application, especially in the wild environment. Specifically, we are interested in cognition-aware computing by integrating environment contexts and multi-modal nonverbal social cues not limited to gaze interaction, body language and para language. More importantly, we extend multi-modal human behavior research to infer the internal states of humans. This is a challenging problem yet important to realize effective interaction between humans and intelligent systems.

It is desirable for intelligent systems like robots, virtual agents, human-machine interfaces to collaborate and interact seamlessly with humans in the era of Industry 5.0, where intelligent systems must work alongside humans to perform a variety of tasks anywhere at home, factories, offices, transit, etc. The underlying technologies to achieve efficient and intelligent collaboration between humans and ubiquitous intelligent systems can be realized by cooperative intelligence, spanning interdisciplinary studies between robotics, AI, human-robot and -computer interaction, computer vision, cognitive science, etc.

One of the main considerations to achieve cooperative intelligence between humans and intelligent systems is to enable everyone and everything to know each other well, like how humans can trust or infer the implicit internal states like intention, emotion, and cognitive states of each other. The importance of empathy to facilitate human-robot interaction has been highlighted in previous studies . However, it is difficult for intelligent systems to estimate the internal states of humans because they are dependent on the complex social dynamics and environment contexts. This requires intelligent systems to be capable of sensing the multi-modal inputs, reasoning the underlying abstract knowledge, and generating the corresponding responses to collaborate and interact with humans.

There are many studies on estimating internal states of humans through measurements of wearables and non-invasive sensors, but it would be difficult to implement these solutions in the wild because of the additional sensors to be worn by humans. One promising solution is to use audiovisual data like nonverbal behavior cues consisting of gaze interaction, facial expression, body language and paralanguage to infer the internal states of humans. Researchers in cognitive and social psychology have long advocated that these nonverbal behaviors are subconsciously generated by humans and reflect the internal states of humans under different contexts. Some salient examples are the studies on emotion recognition using facial and body language in controlled environment. It remains an open question for intelligent systems to sense and recognize nonverbal cues and reason the rich underlying internal states of humans in the wild and noisy environments.

Organizers

Jouh Yeong Chew

Honda Research Institute Japan

jouhyeong.chew@jp.honda-ri.com
Xucong Zhang

TU Delft

xucong.zhang@tudelft.nl
Iolanda Leite

KTH Royal Institute of Technology

iolanda@kth.se
Daisuke Kurabayashi

Tokyo Institute of Technology

kurabayashi.d.aa@m.titech.ac.jp
Eiichi Yoshida

Tokyo University of Science

eiichi.yoshida@rs.tus.ac.jp
Siyu Tang

ETH Zürich

siyu.tang@inf.ethz.ch
Andreas Bulling

University of Stuttgart

andreas.bulling@vis.uni-stuttgart.de