1 code implementation • 16 May 2024 • Yihong Liu, Chunlan Ma, Haotian Ye, Hinrich Schütze
We applied TransMI to three recent strong mPLMs, and our experiments demonstrate that TransMI not only preserves their ability to handle non-transliterated data, but also enables the models to effectively process transliterated data: the results show a consistent improvement of 3% to 34%, varying across different models and tasks.
1 code implementation • 12 Jan 2024 • Yihong Liu, Chunlan Ma, Haotian Ye, Hinrich Schütze
As a consequence, mPLMs are faced with a script barrier: representations from different scripts are located in different subspaces, which can result in crosslingual transfer involving languages of different scripts performing suboptimally.
no code implementations • 9 Jan 2024 • Haotian Ye, Yihong Liu, Chunlan Ma, Hinrich Schütze
In this paper, we introduce MoSECroT Model Stitching with Static Word Embeddings for Crosslingual Zero-shot Transfer), a novel and challenging task that is especially relevant to low-resource languages for which static word embeddings are available.
1 code implementation • 15 Nov 2023 • Yihong Liu, Peiqin Lin, Mingyang Wang, Hinrich Schütze
Instead of pretraining multilingual language models from scratch, a more efficient method is to adapt existing pretrained language models (PLMs) to new languages via vocabulary extension and continued pretraining.
1 code implementation • 26 May 2023 • Yihong Liu, Alexandra Chronopoulou, Hinrich Schütze, Alexander Fraser
By conducting extensive experiments on different language pairs, including similar and distant, high and low-resource languages, we find that our method alleviates the copying problem, thus improving the translation performance on low-resource languages.
no code implementations • 22 May 2023 • Haotian Ye, Yihong Liu, Hinrich Schütze
An interesting line of research in natural language processing (NLP) aims to incorporate linguistic typology to bridge linguistic diversity and assist the research of low-resource languages.
2 code implementations • 22 May 2023 • Yihong Liu, Haotian Ye, Leonie Weissweiler, Renhao Pei, Hinrich Schütze
ColexNet's nodes are concepts and its edges are colexifications.
3 code implementations • 15 May 2023 • Yihong Liu, Haotian Ye, Leonie Weissweiler, Philipp Wicke, Renhao Pei, Robert Zangenfeind, Hinrich Schütze
The resulting measure for the conceptual similarity of two languages is complementary to standard genealogical, typological, and surface similarity measures.
no code implementations • ACL 2022 • Yihong Liu, Haris Jabbar, Hinrich Schütze
The primary novelties of our model are: (a) capturing language-specific sentence representations separately for each language using normalizing flows and (b) using a simple transformation of these latent representations for translating from one language to another.
no code implementations • 31 Oct 2021 • Shubo Yang, Han Han, Yihong Liu, Weisi Guo, Zhibo Pang, Lei Zhang
In this paper, for mmWave secret key generation of physical layer security, we use a reconfigurable intelligent surface (RIS) to induce randomness directly in wireless environments, without adding complexity to transceivers.