Search Results for author: Haixia Han

Found 4 papers, 0 papers with code

SED: Self-Evaluation Decoding Enhances Large Language Models for Better Generation

no code implementations26 May 2024 Ziqin Luo, Haixia Han, Haokun Zhao, Guochao Jiang, Chengyu Du, Tingyun Li, Jiaqing Liang, Deqing Yang, Yanghua Xiao

Existing Large Language Models (LLMs) generate text through unidirectional autoregressive decoding methods to respond to various user queries.

Enhancing Confidence Expression in Large Language Models Through Learning from Past Experience

no code implementations16 Apr 2024 Haixia Han, Tingyun Li, Shisong Chen, Jie Shi, Chengyu Du, Yanghua Xiao, Jiaqing Liang, Xin Lin

Specifically, we first identify three key problems: (1) How to capture the inherent confidence of the LLM?

Large Language Model Can Continue Evolving From Mistakes

no code implementations11 Apr 2024 Haokun Zhao, Haixia Han, Jie Shi, Chengyu Du, Jiaqing Liang, Yanghua Xiao

As world knowledge evolves and new task paradigms emerge, Large Language Models (LLMs) often fall short of meeting new demands due to knowledge deficiencies and outdated information.

Continual Learning Continual Pretraining +4

Small Language Model Can Self-correct

no code implementations14 Jan 2024 Haixia Han, Jiaqing Liang, Jie Shi, Qianyu He, Yanghua Xiao

In this paper, we introduce the \underline{I}ntrinsic \underline{S}elf-\underline{C}orrection (ISC) in generative language models, aiming to correct the initial output of LMs in a self-triggered manner, even for those small LMs with 6 billion parameters.

Language Modelling

Cannot find the paper you are looking for? You can Submit a new open access paper.