no code implementations • 21 May 2024 • Soumyadip Ghosh, Yingdong Lu, Tomasz Nowicki
We study convergence rates of Hamiltonian Monte Carlo (HMC) algorithms with leapfrog integration under mild conditions on stochastic gradient oracle for the target distribution (SGHMC).
no code implementations • 20 Oct 2022 • Soumyadip Ghosh, Yingdong Lu, Tomasz Nowicki, Edith Zhang
We present a framework to analyze MFVI algorithms, which is inspired by a similar development for general variational Bayesian formulations.
no code implementations • 4 Feb 2022 • Soumyadip Ghosh, Yingdong Lu, Tomasz J. Nowicki
We study the convergence of a random iterative sequence of a family of operators on infinite dimensional Hilbert spaces, inspired by the Stochastic Gradient Descent (SGD) algorithm in the case of the noiseless regression, as studied in [1].
no code implementations • 21 Oct 2021 • Soumyadip Ghosh, Yingdong Lu, Tomasz Nowicki
Existing rigorous convergence guarantees for the Hamiltonian Monte Carlo (HMC) algorithm use Gaussian auxiliary momentum variables, which are crucially symmetrically distributed.
no code implementations • 4 Feb 2021 • Soumyadip Ghosh, Yingdong Lu, Tomasz Nowicki
The main purpose of this paper is to facilitate the communication between the Analytic, Probabilistic and Algorithmic communities.
no code implementations • 21 Jan 2021 • Soumyadip Ghosh, Yingdong Lu, Tomasz Nowicki
We establish $L_q$ convergence for Hamiltonian Monte Carlo algorithms.
no code implementations • NeurIPS 2019 • Yingdong Lu, Mark Squillante, Chai Wah Wu
We consider a new family of stochastic operators for reinforcement learning with the goal of alleviating negative effects and becoming more robust to approximation or estimation errors.
no code implementations • 28 May 2019 • Yingdong Lu, Mark S. Squillante, Chai Wah Wu
We consider a new form of reinforcement learning (RL) that is based on opportunities to directly learn the optimal control policy and a general Markov decision process (MDP) framework devised to support these opportunities.
no code implementations • 21 May 2018 • Yingdong Lu, Mark S. Squillante, Chai Wah Wu
We consider a new family of operators for reinforcement learning with the goal of alleviating the negative effects and becoming more robust to approximation or estimation errors.