Search Results for author: Joongkyu Lee

Found 3 papers, 0 papers with code

Randomized Exploration for Reinforcement Learning with Multinomial Logistic Function Approximation

no code implementations30 May 2024 Wooseong Cho, TaeHyun Hwang, Joongkyu Lee, Min-hwan Oh

For our first algorithm, $\texttt{RRL-MNL}$, we adapt optimistic sampling to ensure the optimism of the estimated value function with sufficient frequency and establish that $\texttt{RRL-MNL}$ is both statistically and computationally efficient, achieving a $\tilde{O}(\kappa^{-1} d^{\frac{3}{2}} H^{\frac{3}{2}} \sqrt{T})$ frequentist regret bound with constant-time computational cost per episode.

Nearly Minimax Optimal Regret for Multinomial Logistic Bandit

no code implementations16 May 2024 Joongkyu Lee, Min-hwan Oh

To the best of our knowledge, this is the first work in the contextual MNL bandit literature to prove minimax optimality -- for either uniform or non-uniform reward setting -- and to propose a computationally efficient algorithm that achieves this optimality up to logarithmic factors.

Learning Uncertainty-Aware Temporally-Extended Actions

no code implementations8 Feb 2024 Joongkyu Lee, Seung Joon Park, Yunhao Tang, Min-hwan Oh

In reinforcement learning, temporal abstraction in the action space, exemplified by action repetition, is a technique to facilitate policy learning through extended actions.

Cannot find the paper you are looking for? You can Submit a new open access paper.