Overcoming overfitting in reinforcement learning via Gaussian Process Diffusion Policy

Horprasert, A., Apriaskar, E., Liu, X. et al. (2 more authors) (2025) Overcoming overfitting in reinforcement learning via Gaussian Process Diffusion Policy. In: Proceedings of the 2025 IEEE Statistical Signal Processing Workshop (SSP). 2025 IEEE Statistical Signal Processing Workshop (SSP), 08-11 Jun 2025, Edinburgh, United Kingdom. Institute of Electrical and Electronics Engineers (IEEE) ISBN 9798331518011

Abstract

Metadata

Item Type: Proceedings Paper
Authors/Creators:
Copyright, Publisher and Additional Information:

© 2025 The Authors. Except as otherwise noted, this author-accepted version of a journal article published in Proceedings of the 2025 IEEE Statistical Signal Processing Workshop (SSP) is made available via the University of Sheffield Research Publications and Copyright Policy under the terms of the Creative Commons Attribution 4.0 International License (CC-BY 4.0), which permits unrestricted use, distribution and reproduction in any medium, provided the original work is properly cited. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/

Keywords: Reinforcement Learning; Gaussian Process Regression; Diffusion Policy; OpenAI Gym; Walker2d
Dates:
  • Accepted: 5 April 2025
  • Published (online): 16 July 2025
  • Published: 16 July 2025
Institution: The University of Sheffield
Academic Units: The University of Sheffield > Faculty of Engineering (Sheffield) > School of Electrical and Electronic Engineering
Funding Information:
Funder
Grant number
ENGINEERING AND PHYSICAL SCIENCE RESEARCH COUNCIL
EP/V026747/1
Depositing User: Symplectic Sheffield
Date Deposited: 09 May 2025 15:42
Last Modified: 22 Jul 2025 13:53
Status: Published
Publisher: Institute of Electrical and Electronics Engineers (IEEE)
Refereed: Yes
Identification Number: 10.1109/SSP64130.2025.11073292
Related URLs:
Open Archives Initiative ID (OAI ID):

Export

Statistics