In this paper, we will deal with a linear quadratic optimal control problem with unknown dynamics. As a modeling assumption, we will suppose that the knowledge that an agent has on the current system is represented by a probability distribution π on the space of matrices. Furthermore, we will assume that such a probability measure is opportunely updated to take into account the increased experience that the agent obtains while exploring the environment, approximating with increasing accuracy the underlying dynamics. Under these assumptions, we will show that the optimal control obtained by solving the “average” linear quadratic optimal control problem with respect to a certain π converges to the optimal control driven related to the linear quadratic optimal control problem governed by the actual, underlying dynamics. This approach is closely related to model-based reinforcement learning algorithms where prior and posterior probability distributions describing the knowledge on the uncertain system are recursively updated. In the last section, we will show a numerical test that confirms the theoretical results.
|Titolo:||Convergence results for an averaged LQR problem with applications to reinforcement learning|
PALLADINO, MICHELE (Corresponding)
|Data di pubblicazione:||2021|
|Appare nelle tipologie:||1.1 Articolo in rivista|
File in questo prodotto:
|Pesare2021_Article_ConvergenceResultsForAnAverage.pdf||Documento in Versione Editoriale||Open Access Visualizza/Apri|