GPI treats demonstrations as geometric curves that induces distance and flow fields for simple, efficient, flexible and interpretable imitation learning.
From GPI we obtain multimodal behaviors, better performance, and inference times that are 20–100× faster than diffusion-based policies, while cutting memory usage by orders of magnitude.
Two representative robot experiments highlight GPI coping with unknown complex dynamics and adaptive interaction.
Imitation ultimately means following the expert’s behavior while staying as close as possible. GPI makes this explicit by decoupling imitation into metric learning and policy synthesis.
GPI only needs a meaningful distance—Euclidean, geodesic, cosine, or latent—so we can swap in task-specific encoders, self-supervised autoencoders, or pretrained vision backbones without ever touching the controller.
Progression and attraction flows superposes a stable dynamical system in the actuated space, yielding efficient, interpretable policies.
Simulation results show how GPI generalises across PushT, Robomimic, and Adroit Hand benchmarks.
@misc{GPI,
Author = {Yiming Li and Nael Darwiche and Amirreza Razmjoo and Sichao Liu and Yilun Du and Auke Ijspeert and Sylvain Calinon},
Title = {Geometry-aware Policy Imitation},
Year = {2025},
Eprint = {arXiv:2510.08787},
}