Search Results for author: Wen-Pu Cai

Found 2 papers, 0 papers with code

LCQ: Low-Rank Codebook based Quantization for Large Language Models

no code implementations31 May 2024 Wen-Pu Cai, Wu-Jun Li

Weight quantization has been widely used for model compression, which can reduce both storage and computational cost.

Model Compression Quantization

Weight Normalization based Quantization for Deep Neural Network Compression

no code implementations1 Jul 2019 Wen-Pu Cai, Wu-Jun Li

WNQ adopts weight normalization to avoid the long-tail distribution of network weights and subsequently reduces the quantization error.

Neural Network Compression Quantization

Cannot find the paper you are looking for? You can Submit a new open access paper.