1 code implementation • AKBC 2021 • Thy Thy Tran, Phong Le, Sophia Ananiadou
Unfortunately, both annotation methodologies are costly and time-consuming since they depend on intensive human labour for annotation or for knowledge base creation.
no code implementations • COLING 2020 • Phong Le, Willem Zuidema
Interpreting the inner workings of neural models is a key step in ensuring the robustness and trustworthiness of the models, but work on neural network interpretability typically faces a trade-off: either the models are too constrained to be very useful, or the solutions found by the models are too complex to interpret.
1 code implementation • ACL 2020 • Thy Thy Tran, Phong Le, Sophia Ananiadou
Unsupervised relation extraction (URE) extracts relations between named entities from raw text without manually-labelled data and existing knowledge bases (KBs).
1 code implementation • ACL 2019 • Phong Le, Ivan Titov
First, we construct a high recall list of candidate entities for each mention in an unlabeled document.
Ranked #17 on Entity Disambiguation on AIDA-CoNLL
1 code implementation • ACL 2019 • Phong Le, Ivan Titov
As the learning signal is weak and our surrogate labels are noisy, we introduce a noise detection component in our model: it lets the model detect and disregard examples which are likely to be noisy.
2 code implementations • ACL 2018 • Phong Le, Ivan Titov
Entity linking involves aligning textual mentions of named entities to their corresponding entries in a knowledge base.
1 code implementation • CONLL 2017 • Phong Le, Ivan Titov
Coreference evaluation metrics are hard to optimize directly as they are non-differentiable functions, not easily decomposable into elementary decisions.
no code implementations • WS 2016 • Phong Le, Marc Dymetman, Jean-Michel Renders
We introduce an LSTM-based method for dynamically integrating several word-prediction experts to obtain a conditional language model which can be good simultaneously at several subtasks.
no code implementations • WS 2016 • Phong Le, Willem Zuidema
Recursive neural networks (RNN) and their recently proposed extension recursive long short term memory networks (RLSTM) are models that compute representations for sentences, by recursively combining word embeddings according to an externally provided parse tree.
no code implementations • HLT 2015 • Phong Le, Willem Zuidema
We present a self-training approach to unsupervised dependency parsing that reuses existing supervised and unsupervised parsing algorithms.
Ranked #1 on Unsupervised Dependency Parsing on Penn Treebank
1 code implementation • SEMEVAL 2015 • Phong Le, Willem Zuidema
We are proposing an extension of the recursive neural network that makes use of a variant of the long short-term memory architecture.