Within the last decade, advances in deep learning, coupled with the creation of large, freely available datasets (e.g., ImageNet), have resulted in remarkable progress in the computer vision, NLP, and broader AI communities. This progress has enabled models to begin to obtain superhuman performance on a wide variety of passive tasks. However, this progress has also enabled a paradigm shift that a growing collection of researchers take aim at: the creation of an embodied agent (e.g., a robot) which learns, through interaction and exploration, to creatively solve challenging tasks within its environment.
The goal of this workshop is to bring together researchers from the fields of computer vision, language, graphics, and robotics to share and discuss the current state of intelligent agents that can:
The Embodied AI 2021 workshop will be held virtually in conjunction with CVPR 2021. It will feature a host of invited talks covering a variety of topics in Embodied AI, many exciting challenges, a poster session, and panel discussions.
Date. June 20th, 11 AM PST.
Panel. The panel consists of speakers at this workshop.
Moderator. Erik Wijmans.
Topics. The topics are based on questions, likely involving cognitive development in humans, progress in embodied AI tasks, sim-2-real transfer, robotics, embodied AI for all, and more!
Date. June 20th, 3 PM PST.
Panel. The panel consists of challenge organizers who organized navigation tasks.
Moderator. Luca Weihs.
Topics. The topics are based on questions, likely involving navigation benchmarks and tasks, the "reality" gap, robotics, simulation platforms, and more!
Date. June 20th, 5 PM PST.
Panel. The panel consists of challenge organizers who organized interaction tasks.
Moderator. Chengshu (Eric) Li.
Topics. The topics are based on questions, likely involving interaction benchmarks and tasks, vision-and-language, rearrangement, leveraging audio, the "reality" gap, robotics, simulation platforms, and more!
The Embodied AI 2021 workshop is hosting many exciting challenges covering a wide range of topics such as rearrangement, visual navigation, vision-and-language, and audio-visual navigation. More details regarding data, submission instructions, and timelines can be found on the individual challenge websites.
Challenge winners will be given the opportunity to present a talk at the workshop. Since many challenges can be grouped into similar tasks, we encourage participants to submit models to more than 1 challenge. The table below describes, compares, and links each challenge.
We invite high-quality 2-page extended abstracts in relevant areas, such as:
Note. The order of the papers is randomized each time the page is refreshed.