no code implementations • 23 Apr 2024 • Constanza Fierro, Jiaang Li, Anders Søgaard
The purpose of instruction tuning is enabling zero-shot performance, but instruction tuning has also been shown to improve chain-of-thought reasoning and value alignment (Si et al., 2023).
no code implementations • 18 Mar 2024 • Qinghua Zhao, Jiaang Li, Lei LI, Zenghui Zhou, Junfeng Liu
Existing works have studied the impacts of the order of words within natural text.
no code implementations • 8 Feb 2024 • Yong Cao, Wenyan Li, Jiaang Li, Yifei Yuan, Antonia Karamolegkou, Daniel Hershcovich
Pretrained large Vision-Language models have drawn considerable interest in recent years due to their remarkable performance.
1 code implementation • 24 Oct 2023 • Jiaang Li, Quan Wang, Yi Liu, Licheng Zhang, Zhendong Mao
We analyze this phenomenon and reveal that entity codes, the quantization outcomes for expressing entities, have higher entropy at the code level and Jaccard distance at the codeword level under random entity quantization.
1 code implementation • 20 Oct 2023 • Antonia Karamolegkou, Jiaang Li, Li Zhou, Anders Søgaard
Language models may memorize more than just facts, including entire chunks of texts seen during training.
no code implementations • 5 Jun 2023 • Laura Cabello, Jiaang Li, Ilias Chalkidis
We then evaluate its ability to acquire new knowledge and include it in its reasoning process.
1 code implementation • 2 Jun 2023 • Jiaang Li, Antonia Karamolegkou, Yova Kementchedjhieva, Mostafa Abdou, Sune Lehmann, Anders Søgaard
Human language processing is also opaque, but neural response measurements can provide (noisy) recordings of activation during listening or reading, from which we can extract similar representations of words and phrases.
1 code implementation • 1 Apr 2023 • Jiaang Li, Quan Wang, Zhendong Mao
Relation prediction on knowledge graphs (KGs) is a key research topic.
no code implementations • 13 Feb 2023 • Jiaang Li, Yova Kementchedjhieva, Anders Søgaard
Large-scale pretrained language models (LMs) are said to ``lack the ability to connect [their] utterances to the world'' (Bender and Koller, 2020).