Publication 2025

Flow-based Domain Randomization for Learning and Sequencing Robotic Skills

Two robots learning a high-dimensional bimanual insertion task under flow-based domain randomization.

Abstract

Domain randomization in reinforcement learning is an established technique for increasing the robustness of control policies trained in simulation. By randomizing environment properties during training, the learned policy can become robust to uncertainties along the randomized dimensions. While the environment distribution is typically specified by hand, in this paper we investigate automatically discovering a sampling distribution via entropy-regularized reward maximization of a normalizing-flow–based neural sampling distribution. We show that this architecture is more flexible and provides greater robustness than existing approaches that learn simpler, parameterized sampling distributions, as demonstrated in six simulated and one real-world robotics domain. Lastly, we explore how these learned sampling distributions, along with a privileged value function, can be used for out-of-distribution detection in an uncertainty-aware multi-step manipulation planner.

Download publication

Associated Researchers

Aidan Curtis

MIT CSAIL

Eric Li

MIT CSAIL

Michael Noseworthy

MIT CSAIL

Nishad Gothoskar

MIT CSAIL

Leslie Pack Kaelbling

MIT CSAIL

Get in touch

Something pique your interest? Get in touch if you’d like to learn more about Autodesk Research, our projects, people, and potential collaboration opportunities.

Contact us