Search Results for author: Michal Nauman

Found 5 papers, 2 papers with code

A Case for Validation Buffer in Pessimistic Actor-Critic

no code implementations1 Mar 2024 Michal Nauman, Mateusz Ostaszewski, Marek Cygan

VPL uses a small validation buffer to adjust the levels of pessimism throughout the agent training, with the pessimism set such that the approximation error of the critic targets is minimized.

On the Theory of Risk-Aware Agents: Bridging Actor-Critic and Economics

no code implementations30 Oct 2023 Michal Nauman, Marek Cygan

Risk-aware Reinforcement Learning (RL) algorithms like SAC and TD3 were shown empirically to outperform their risk-neutral counterparts in a variety of continuous-action tasks.

Reinforcement Learning (RL)

On Many-Actions Policy Gradient

1 code implementation24 Oct 2022 Michal Nauman, Marek Cygan

We study the variance of stochastic policy gradients (SPGs) with many action samples per state.

Low-Variance Policy Gradient Estimation with World Models

1 code implementation29 Oct 2020 Michal Nauman, Floris den Hengst

In WMPG, a WM is trained online and used to imagine trajectories.

Cannot find the paper you are looking for? You can Submit a new open access paper.