TEACh Two-Agent Task Completion (TATC) Challenge
Code, precomputed features, and AI2Thor simulator are all available for a quick start on GitHub
This challenge focuses on how to produce instructions for, ask and answer questions about, and carry out embodied visual tasks in a shared virtual environment. We challenge researchers to consider partner models of the world, pragmatics of instruction and question generation, barge-in communication, and instruction following from verbal and nonverbal cues. Such challenges are not captured by current datasets for embodiment [1, 2, 3, 4, 5].
Guidelines
Participants are required to upload their model to our evaluation server (coming soon!) with [EAI22]
in the submission title, e.g., [EAI22] Seq2seq Model. The evaluation server automatically evaluates the models on an unseen test set. Final numbers for the prize challenge will be frozen on Jun 12. Winning submissions will be required to submit a brief (private) report of technical details for validity checking. We will also conduct a quick code inspection to ensure that the challenge rules weren't violated (e.g., peeking at unavailable info on test scenes for either the Commander or Follower agent).
Dataset
The challenge is based on the TEACh Dataset, which contains over 3,000 episodes of human-human dialogues for guiding a human-controlled agent to complete household chores in the AI2THOR simulator. Agents interact with environments through discrete actions with end effector click positions, and with one another via a text-chat interface.
Timeline | |
Challenge Opens | Feb 14 |
Leaderboard closes | Jun 12 |
Winner announcement | Jun 17 |
Evaluation (Links coming soon!)
It is likely that you will submit your pre-trained Commander and Follower agents, which will be run on an evaluation server on unseen data.
Metric
The submissions will be ranked by Unseen Success Rate.
Rules
[EAI22]
in the submission title e.g., [EAI22] Seq2seq Model.