Learning from less: SINDy Surrogates in RL

Research output: Contribution to conferencePaperpeer-review

Abstract

This paper presents a method to create simpler versions of reinforcement learning (RL) environments using the Sparse Identification of Nonlinear Dynamics (SINDy) algorithm. The method was tested in OpenAI Gym's Mountain Car and Lunar Lander environments. Results show that these SINDy-based models can closely match the real environment's behavior while cutting down computational cost by 20–35%. With just 75 data points for Mountain Car and 1000 for Lunar Lander, the models achieved over 0.997 state-wise correlation, with very low prediction errors (as low as 3.11 × 10⁻⁶ for Mountain Car velocity and 1.42 × 10⁻⁶ for Lunar Lander position). RL agents trained in these simplified environments needed fewer steps to learn (65,075 vs. 100,000 for Mountain Car and 801,000 vs. 1,000,000 for Lunar Lander) and performed similarly to those trained in the full environments. This work offers a fast and accurate way to create interpretable models for use in model-based RL.
Original languageEnglish
DOIs
Publication statusPublished - 26 Apr 2025
EventICLR 2025 Workshop on World Models: Understanding, Modelling and Scaling
- , Singapore
Duration: 24 Apr 202528 Apr 2025
https://sites.google.com/view/worldmodel-iclr2025/

Workshop

WorkshopICLR 2025 Workshop on World Models: Understanding, Modelling and Scaling
Abbreviated titleICLR
Country/TerritorySingapore
Period24/04/2528/04/25
Internet address

Keywords

  • Reinforcement Learning (RL)
  • Model Based Learning
  • Non Linear Dynamics

Cite this