Optimized user-guided motion control of modular robots
Transferring motion instructions from a user enables robots to perform new and unforeseen operations. Robot collectives, in particular, offer greater adaptability to changing tasks and environments. However, effectively transferring motion instructions becomes challenging as the collective’s shape and size evolve. These changes often require additional system constraints to maintain robust motion control, which typically depends on pre-programmed knowledge of new tasks, ultimately limiting the collective’s adaptability. To overcome the above challenges, we propose a physical and computational platform for user-guided control of self-reconfigurable modular robots. This platform consists of an optimization scheme for online processing of user commands, which prevents any modular robot actions that would violate system or environment constraints. The second component consists of Joint-space Joysticks, which match the robot’s morphology, enabling the user to control diverse and dynamically changing modular robot structures through direct physical interaction. We present a platform that enables users to safely control modular, shape-changing robots through a physical interface. We demonstrate the platform’s efficacy and generalizability across a diverse set of modular robot morphologies using two independent robotic systems — Mori3 and Roombots — performing a range of tasks including pick-and-place, human assistance, legged locomotion, and workspace expansion.
10.1038_s41467-025-63706-6.pdf
Main Document
Published version
openaccess
CC BY-NC-ND
2.41 MB
Adobe PDF
8745ac02438bc59c147bf122b3167e24