About Human-aware Embodied AI (HEAI) | IROS 2025 Workshop


Embodied AI, i.e., the development of technologies capable of interacting with their physical environment, witnessed substantial progress in the last years, mainly thanks to the advent of Large Language Models (LLMs), Multimodal Large Language Models (MLLMs), and other large foundation AI approaches. These breakthroughs have enabled human users to interact with embodied agents (e.g., robots) through natural language, instructing an embodied agent to navigate to a given position, searching for a specific object, or interacting with the object to perform a certain manipulation.


However, while being inherently human-centric, these problems are often addressed without considering the presence of human users and their interaction with the agent throughout the execution of the embodied task. Considering that the long-term goal of Embodied AI is to create a Co-Habitat for humans and agents, Bi-directional communication (i.e., human-to-agent and agent-to-human) emerges as a key challenge that must be addressed to effectively resolve uncertainty during the full cycle of robotic perception, reasoning, planning, action, and verification.

However, the problem of creating an intuitive, effective, trustworthy, and robust human-agent Co-Habitat remains largely unexplored.


This workshop will address the challenges of creating seamless human-agent Co-Habitats, focusing on trustworthiness (when agents don't know), robustness (when agents ask for help), and intuitive collaboration (bi-directional interaction). We aim to explore methods for integrating MLLMs into embodied agents for effective multimodal interactions, identifying and mitigating the risks of model hallucinations, and modeling human error and communication patterns.


Schedule

Tentative schedule as follows:


Workshop Schedule (25 min talk + 5 min Q&A)
Time Talk Comments
8:30 - 8:40 Welcome & Opening Remarkss
8:40 - 9:10 Daniele Pucci ergoCub: the Embodied AI human aware humanoid
9:10 - 9:40 He Wang Toward Human-Level Dexterous Grasping Data Synthesis
9:40 - 10:10 Roozbeh Mottaghi Leveraging Human Experience in Robotics
10:10 - 11:00 Coffee Break, Poster session & AMA (Ask Me Anything) panel
11:00 - 11:25 Best posters presentation 8 minutes each + 3 min Q&A
11:25 - 11:55 Alessandro Vinciarelli Introduction to Social AI for Robotics
11:55 - 12:25 Gaofeng Li Agile Telerobotic Systems: Embodied and Intelligent Robotic Manipulation Learning with Human-in-the-Loop
12:25 - 12-30 Closing Remarks

Call for Contributions

We call for papers covering relevant topics, including but not limited to:


    • Collaborative Human-Agent interaction
    • Multimodal Human-Agent interaction
    • Interaction-enabled simulators / benchmarks
    • Interaction personalization
    • Interaction adaptation
    • Uncertainty understanding
    • Foundation models (VLM/VLAs)
    • Human behaviour modeling
    • Embodied reasoning
    • Ethics & Safety

Submission Guidelines
We invite submissions of up to 4 pages (excluding references) of non-archival papers using the official template. Submissions previously published elsewhere are also welcome.

To submit a paper, please see the official OpenReview page.

Further details about the Poster Presentation will be shared soon.

Organizing Committee

Placeholder image
PhD candidate in the National PhD Programme in AI at the Polytechnic of Turin, Italy.
Placeholder image
Researcher in the DVL unit in Fondazione Bruno Kessler (FBK), Italy.
Placeholder image
Professor at the University of Science and Technology, Beijing, China.
Placeholder image
Research Scientist at FAIR (Meta), Embodied AI team.                                       
Placeholder image
Associate Professor at Simon Fraser University, British Columbia, Canada.
Placeholder image

Alessandro vinciarelli

https://vinciarelli.net/

Full Professor at the University of Glasgow, UK.
Placeholder image
Tenured senior researcher leading the PAVIS at IIT.
Placeholder image
Professor at the University of Verona, Italy.

Sponsored by: