Real-is-Sim: Bridging the Sim-to-Real Gap with a Dynamic Digital Twin for Real-World Robot Policy Evaluation

1Robotics and AI Institute 2Queensland University of Technology

The real-is-sim framework operating with an always-in-the-loop simulator. The learnt PushT policy acts on the simulator and the simulator synchronizes itself with the real world. The real robot mimics the simulated robot.

Abstract

Recent advancements in behavior cloning have enabled robots to perform complex manipulation tasks. However, accurately assessing training performance remains challenging, particularly for real-world applications, as behavior cloning losses often correlate poorly with actual task success. Consequently, researchers resort to success rate metrics derived from costly and time-consuming real-world evaluations, making the identification of optimal policies and detection of overfitting or underfitting impractical. To address these issues, we propose real-is-sim, a novel behavior cloning framework that incorporates a dynamic digital twin (based on Embodied Gaussians) throughout the entire policy development pipeline: data collection, training, and deployment. By continuously aligning the simulated world with the physical world, demonstrations can be collected in the real world with states extracted from the simulator. The simulator enables flexible state representations by rendering image inputs from any viewpoint or extracting low-level state information from objects embodied within the scene. During training, policies can be directly evaluated within the simulator in an offline and highly parallelizable manner. Finally, during deployment, policies are run within the simulator where the real robot directly tracks the simulated robot's joints, effectively decoupling policy execution from real hardware and mitigating traditional domain-transfer challenges. We validate real-is-sim on the PushT manipulation task, demonstrating strong correlation between success rates obtained in the simulator and real-world evaluations.

Framework

Framework illustration

The real-is-sim framework seamlessly transitions between online and offline modes because it uses an always-in-the-loop simulator as a mediator.

Learnt Policies

Policies learnt from different representations extracted from the simulator state.

Online Mode

Our system operating in online mode with Embodied Gaussians acting as a mediator between the policy and the real world.

Offline Mode

Our system operating in offline mode evaluating an imitation learning policy. Offline mode can render virtual cameras for multiple scenes at the same time and evaluate vision based policies as well as state based policies.

Gripper Virtual Camera

An imitation learning policy trained on a virtually rendered image from a virtual camera mounted on the end effector.

Static Virtual Camera

An imitation learning policy trained on a virtually rendered image from a static virtual camera.

Interesting Exploration Behaviour

The imitation learning policy explores the space to find the T-Block when the T-Block is not in view of its virtual gripper camera.

Desynchronization

The imitation learning policy creates actions that actively prevent the visual correction mechanism from resynchronizing the simulator.

Related Links

Embodied Gaussians is the underlying simulator that powers real-is-sim.