Published On Jan 4, 2024
Introduce 𝐌𝐨𝐛𝐢𝐥𝐞 𝐀𝐋𝐎𝐇𝐀🏄 -- Learning!
With 50 demos, our robot can autonomously complete complex mobile manipulation tasks:
- cook and serve shrimp 🦐
- call and take an elevator 🛗
- store a 3 Ibs pot to a two-door cabinet
- push 5 consecutive chairs
- rinse pan using a water faucet
- play high fives with people
Co-led by Tony Z. Zhao, Chelsea Finn
Our robot can consistently handle these tasks, succeeding:
- 9 times in a row for Wipe Wine
- 5 times for Call Elevator
- robust against distractors for Use Cabinet
- extrapolate to chairs unseen during training
How do we achieve this with only 50 demos? The key is to co-train imitation learning algorithms with static ALOHA data. We found this to consistently improve performance, especially for tasks that require precise manipulation.
Co-training (1) improves the performance across all tasks, (2) is compatible with ACT, Diffusion Policy and VINN, (3) is robust to different data mixtures.
We open-source all the software and data of Mobile ALOHA!
Project Website 🛜: https://lnkd.in/gE6A43fR
Code for Imitation Learning 🖥️: https://lnkd.in/gDCmgy_E
Data 📊: https://lnkd.in/gCJJtmvT
#AgileXRobotics #AI #UGV #AGV #Tracer #MobileALOHA