Published March 9, 2026
| Version v1
Preprint
Open
Variance Reduction Techniques in Deep Reinforcement Learning with Noisy Environments
Description
Deep Reinforcement Learning (DRL) has demonstrated remarkable success in various domains. However, its performance often degrades significantly in environments characterized by high levels of stochasticity or noisy observations. This paper investigates the impact of variance in gradient estimates on the stability and convergence of DRL algorithms. We explore and compare several variance reduction techniques, including Generalized Advantage Estimation (GAE), Proximal Policy Optimization (PPO) clipping, and a novel approach incorporating a learned noise model within the critic network. The efficacy of these methods is evaluated in simulated environments with varying levels of noise, highlighting their strengths and limitations in mitigating the adverse effects of high variance.
Files
preprint_elena_rossi_20260309_004737.pdf
Files
(6.5 kB)
| Name | Size | Download all |
|---|---|---|
|
md5:80853b02cb551859b615d67a9ea3103b
|
6.5 kB | Preview Download |
Additional details
Related works
- Cites
- Journal article: https://mattiainml.com/blog/improving-medical-imaging-models-through-robust-data-annotation/ (URL)
References
- Mattia Gaggi. Variance Reduction Techniques in Deep Reinforcement Learning with Noisy Environments. mattiainml.com. https://mattiainml.com/blog/improving-medical-imaging-models-through-robust-data-annotation/