We present ToddleBot's mechatronic design in the figure below. We highlight ToddlerBot's 30 active DoFs with orange markers: 7 DoFs per arm, 6 DoFs per leg, a 2-DoF neck, and a 2-DoF waist. Green markers indicate two end-effector designs—a compliant palm and a parallel-jaw gripper. Purple markers denote the sensor and electronics layout with exploded views, featuring two fisheye cameras, a speaker, two microphones, an IMU, and a Jetson Orin NX computer.
1x Speed unless otherwise noted.
Arm Span Test: With a torso volume of 13x9x12 cm³, ToddlerBot successfully grasps objects 14 times its torso volume (27x24x31 cm³) using its compliant palm gripper.
Payload Test: ToddlerBot lifts 1,484 g (40% of its total weight) while maintaining balance. To determine the limit, screws are incrementally added to a 3D-printed gripping cup until it falls.
Endurance Test: Running a walking RL policy, ToddlerBot lasts 19 minutes before overheating affects stability. It withstands up to 7 falls before breaking, but repairs take only 21 minutes of 3D printing and 14 minutes of assembly for full restoration.
Conversation and Push-ups: This is achieved by integrating OpenAI’s Realtime API with GPT-4o for speech-to-text and text-to-speech. Push-ups are zero-shot sim-to-real transfer of open-loop keyframe animation.
Pull-ups: We use an AprilTag to help ToddlerBot accurately locate the horizontal bar. The rest is a zero-shot sim-to-real transfer of open-loop keyframe animation.
Omnidirectional Walking: ToddlerBot achieves omnidirectional walking with RL and zero-shot sim-to-real.
Bimanual Manipulation: This is an RGB-based diffusion policy trained with 60 demonstrations.
Full-body Manipulation: This is also an RGB-based diffusion policy trained with 60 demonstrations.
Skill Chaining: ToddlerBot first executes a diffusion policy to grasp the handle, while maintaining that pose, switched to the RL policy to push the wagon forward.
Manipulation Policy Zero-Shot Transfer: We successfully transfer manipulation policy trained with data collected on Toddlerbot instance to another.
Two-Instance Collaboration Task: To demonstrate their equivalent performance of two Toddlerbot Instances, both robots collaborate on a long-horizon room tidying task.
Easy-to-Build Test: We show that ToddlerBot is easy to build with fully open-source assembly manuals and videos.
Complete the CAPTCHA to reveal the invite links.
@article{shi2025toddlerbot,
title={ToddlerBot: Open-Source ML-Compatible Humanoid Platform for Loco-Manipulation},
author={Shi, Haochen and Wang, Weizhuo and Song, Shuran and Liu, C. Karen},
journal={arXiv preprint arXiv:2502.00893},
year={2025}
}
The authors would like to express their gratitude to Kaizhe Hu for assembling the second instance of ToddlerBot and assisting with keyframe animation and demo recording. We also extend our thanks to Huy Ha, Yen-Jen Wang, Pei Xu, and Yifan Hou for their insightful discussions on locomotion, and to Sirui Chen, Chen Wang, and Yunfan Jiang for valuable input on manipulation policy deployment. We are grateful to Albert Wu for his guidance on mathematical formulation and notation. Additionally, we thank João Pedro Araújo for his assistance with the motion capture system. Finally, we appreciate the helpful discussions from all members of TML and REALab. This work was supported by National Science Foundation NSF-FRR-2153854, NSF-2143601, NSF-2037101, Sloan Fellowship, Stanford Institute for Human-Centered Artificial Intelligence, and Stanford Wu Tsai Human Performance Alliance.