Interpretable Function Approximation with Gaussian Processes in Value-Based Model-Free Reinforcement Learning

Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

51 Downloads (Pure)

Abstract

Estimating value functions in Reinforcement Learning (RL) for continuous spaces is challenging. While traditional function approximators, such as linear models, offer interpretability, they are limited in their representation capabilities. In contrast, Deep Neural Networks (DNN) can model more complex functions but are less interpretable. Gaussian Process (GP) models bridge this gap by offering interpretable uncertainty estimates while modelling complex nonlinear functions. This work introduces a Bayesian nonparametric framework using GPs, including Sparse Variational (SVGP) and Deep GPs (DGP), for off-policy and on-policy learning. Results on popular classic control environments show that SVGPs/DGPs outperform linear models but converge slower than their DNN counterparts.

Original languageEnglish
Title of host publicationProceedings of the Northern Lights Deep Learning Workshop 2025
PublisherML Research Press
Pages141-154
Number of pages14
Publication statusPublished - 2025
Event6th Northern Lights Deep Learning Conference, NLDL 2025 - Tromso, Norway
Duration: 7-Jan-20259-Jan-2025

Publication series

NameProceedings of Machine Learning Research
Volume265
ISSN (Print)2640-3498

Conference

Conference6th Northern Lights Deep Learning Conference, NLDL 2025
Country/TerritoryNorway
CityTromso
Period07/01/202509/01/2025

Fingerprint

Dive into the research topics of 'Interpretable Function Approximation with Gaussian Processes in Value-Based Model-Free Reinforcement Learning'. Together they form a unique fingerprint.

Cite this