Search Results for author: Shangyu Xing

Found 3 papers, 1 papers with code

AlignGPT: Multi-modal Large Language Models with Adaptive Alignment Capability

no code implementations23 May 2024 Fei Zhao, Taotian Pang, Chunhui Li, Zhen Wu, Junjie Guo, Shangyu Xing, Xinyu Dai

In the pre-training stage, instead of treating all image-text pairs equally, we assign different levels of alignment capabilities to different image-text pairs.

Language Modelling Large Language Model

EFUF: Efficient Fine-grained Unlearning Framework for Mitigating Hallucinations in Multimodal Large Language Models

no code implementations15 Feb 2024 Shangyu Xing, Fei Zhao, Zhen Wu, Tuo An, WeiHao Chen, Chunhui Li, Jianbing Zhang, Xinyu Dai

Multimodal large language models (MLLMs) have attracted increasing attention in the past few years, but they may still generate descriptions that include objects not present in the corresponding images, a phenomenon known as object hallucination.

Hallucination

DRIN: Dynamic Relation Interactive Network for Multimodal Entity Linking

1 code implementation9 Oct 2023 Shangyu Xing, Fei Zhao, Zhen Wu, Chunhui Li, Jianbing Zhang, Xinyu Dai

Multimodal Entity Linking (MEL) is a task that aims to link ambiguous mentions within multimodal contexts to referential entities in a multimodal knowledge base.

Entity Linking Relation

Cannot find the paper you are looking for? You can Submit a new open access paper.