A preview of this full-text is provided by Springer Nature.
Content available from Machine Learning
This content is subject to copyright. Terms and conditions apply.
Vol.:(0123456789)
Machine Learning (2025) 114:4
https://doi.org/10.1007/s10994-024-06724-7
Qualitative control learning can be much faster
thanreinforcement learning
DomenŠoberl1· IvanBratko2
Received: 5 April 2024 / Revised: 20 August 2024 / Accepted: 27 September 2024 /
Published online: 14 January 2025
© The Author(s), under exclusive licence to Springer Science+Business Media LLC, part of Springer Nature 2025
Abstract
Reinforcement learning has emerged as a prominent method for controlling dynamic sys-
tems in the absence of a precise mathematical model. However, its reliance on extensive
interactions with the environment often leads to prolonged training periods. In this paper,
we propose an alternative approach to learning control policies that focuses on learning
qualitative models and uses symbolic planning to derive a qualitative plan for the control
task, which is executed by an adaptive reactive controller. We conduct experiments utiliz-
ing our approach on the cart-pole problem, a standard benchmark in dynamic system con-
trol. We additionally extend this problem domain to include uneven terrains, such as driv-
ing over craters or hills, to assess the robustness of learned controllers. Our results indicate
that qualitative learning offers significant advantages over reinforcement learning in terms
of sample efficiency, transferability, and interpretability. We demonstrate that our proposed
approach is at least two orders of magnitude more sample efficient in the cart-pole domain
than the usual variants of reinforcement learning.
Keywords Qualitative modeling· Qualitative reasoning· Qualitative control· Transfer
learning
Editors:Rita P.Ribeiro, Ana Carolina Lorena and Albert Bifet.
* Domen Šoberl
domen.soberl@famnit.upr.si
Ivan Bratko
ivan.bratko@fri.uni-lj.si
1 Department ofInformation Sciences andTechnologies, Faculty ofMathematics, Natural Sciences
andInformation Technologies, University ofPrimorska, Glagoljaška 8, 6000Koper, Slovenia
2 Artificial Intelligence Laboratory, Faculty ofComputer andInformation Science, University
ofLjubljana, Večna pot 113, 1000Ljubljana, Slovenia
Content courtesy of Springer Nature, terms of use apply. Rights reserved.