Search Results for author: Haoze He

Found 4 papers, 0 papers with code

Sparse Matrix in Large Language Model Fine-tuning

no code implementations24 May 2024 Haoze He, Juncheng Billy Li, Xuan Jiang, Heather Miller

In this work, we introduce a method for selecting sparse sub-matrices that aim to minimize the performance gap between PEFT vs. full fine-tuning (FT) while also reducing both fine-tuning computational cost and memory cost.

Adjacent Leader Decentralized Stochastic Gradient Descent

no code implementations18 May 2024 Haoze He, Jing Wang, Anna Choromanska

This work focuses on the decentralized deep learning optimization framework.

Accelerating Parallel Stochastic Gradient Descent via Non-blocking Mini-batches

no code implementations2 Nov 2022 Haoze He, Parijat Dube

In this paper, we propose the (de)centralized Non-blocking SGD (Non-blocking SGD) which can address the straggler problem in a heterogeneous environment.

Blocking Computational Efficiency

RCD-SGD: Resource-Constrained Distributed SGD in Heterogeneous Environment via Submodular Partitioning

no code implementations2 Nov 2022 Haoze He, Parijat Dube

The convergence of SGD based distributed training algorithms is tied to the data distribution across workers.

Cannot find the paper you are looking for? You can Submit a new open access paper.