π Participate in the Challenge
Thanks for your interest in the Embodied Agent Interface Challenge!
Follow the instructions below to get started with the Development Phase of the EAI Challenge, and most importantly, have fun!
π Resources
To help you get up to speed and make the most of the EAI Challenge, we have prepared a set of essential resources. We recommend exploring them in the following order for the smoothest experience:
-
πTutorial: Step-by-step guide to setting up your environment and understanding the challenge
-
πDocumentation: Complete reference for evaluating and troubleshooting four ability modules
-
π³Docker Image: Prebuilt environment for running your experiments hassle-free
π¦ Submission Preparation
To make participation more accessible to the broader embodied AI community, we have designed a straightforward submission process. You are not required to set up or run the complex BEHAVIOR or VirtualHome simulation environments. All you need to provide is your modelβs outputs. We will take care of the rest and perform evaluation internally in a sandboxed environment within our infrastructure.
We have created a Starter Kit with all the essentials to help you get started quickly. The kit includes:
llm_prompts/
: A directory containing all the prompts you will use to query your model.sample_submission/
: A sample submission folder that shows the required format and structure for your modelβs outputs.starter.ipynb
: A notebook to guide you through the process of generating outputs and preparing your submission.
π Starter Kit Structure
eai_starter_kit/
βββ llm_prompts/
β βββ behavior_action_sequencing_prompts.json
β βββ behavior_goal_interpretation_prompts.json
β βββ behavior_subgoal_decomposition_prompts.json
β βββ behavior_transition_modeling_prompts.json
β βββ virtualhome_action_sequencing_prompts.json
β βββ virtualhome_goal_interpretation_prompts.json
β βββ virtualhome_subgoal_decomposition_prompts.json
β βββ virtualhome_transition_modeling_prompts.json
βββ sample_submission/
β βββ behavior_action_sequencing_outputs.json
β βββ behavior_goal_interpretation_outputs.json
β βββ behavior_subgoal_decomposition_outputs.json
β βββ behavior_transition_modeling_outputs.json
β βββ virtualhome_action_sequencing_outputs.json
β βββ virtualhome_goal_interpretation_outputs.json
β βββ virtualhome_subgoal_decomposition_outputs.json
β βββ virtualhome_transition_modeling_outputs.json
βββ starter.ipynb
π€ EvalAI Submission
To submit your modelβs outputs for evaluation, please follow these steps:
-
Review the EvalAI documentation: Familiarize yourself with the submission process and officially join the EAI Challenge. While at the same time, please fill out the Participant Information Form to help us better organize the competition.
-
Prepare Your Submission: Organize your modelβs outputs according to the structure outlined in the sample_submission in the Starter Kit. Please make sure names of 8 required output files exactly match the expected names as shown in the sample_submission.
-
Upload to EvalAI: Use the EvalAI platform to upload your submission. Please zip your submission folder before uploading. The zipped folder should contain all required output files in the correct structure as shown below.
sample_submission.zip/ βββ behavior_action_sequencing_outputs.json βββ behavior_goal_interpretation_outputs.json βββ behavior_subgoal_decomposition_outputs.json βββ behavior_transition_modeling_outputs.json βββ virtualhome_action_sequencing_outputs.json βββ virtualhome_goal_interpretation_outputs.json βββ virtualhome_subgoal_decomposition_outputs.json βββ virtualhome_transition_modeling_outputs.json
For Mac users, you can use the following command in the terminal to zip your submission folder without including unnecessary hidden files like
.DS_Store
or__MACOSX
:zip -r sample_submission.zip sample_submission -x "*.DS_Store" -x "__MACOSX/*"
A sample submission should look like this:
-
Monitor Your Submission: After submitting, you can monitor the submission status on My Submissions page and performance of your submission on the Leaderboard. Please note that because the evaluation involves interacting with the simulation environments, it may take 10β30 minutes to receive your results.
We look forward to seeing your innovative solutions in action! If you have any questions or need assistance, donβt hesitate to reach out to us at TianweiBao@u.northwestern.edu
or post in our Slack.