no code implementations • WMT (EMNLP) 2021 • Viktor Hangya, Qianchu Liu, Dario Stojanovski, Alexander Fraser, Anna Korhonen
The performance of NMT systems has improved drastically in the past few years but the translation of multi-sense words still poses a challenge.
no code implementations • Findings (EMNLP) 2021 • Alan Ansell, Edoardo Maria Ponti, Jonas Pfeiffer, Sebastian Ruder, Goran Glavaš, Ivan Vulić, Anna Korhonen
While offering (1) improved fine-tuning efficiency (by a factor of around 50 in our experiments), (2) a smaller parameter budget, and (3) increased language coverage, MAD-G remains competitive with more expensive methods for language-specific adapter training across the board.
1 code implementation • NAACL 2022 • Marinela Parović, Goran Glavaš, Ivan Vulić, Anna Korhonen
Adapter modules enable modular and efficient zero-shot cross-lingual transfer, where current state-of-the-art adapter-based approaches learn specialized language adapters (LAs) for individual languages.
no code implementations • Findings (ACL) 2022 • Evgeniia Razumovskaia, Ivan Vulić, Anna Korhonen
Scaling dialogue systems to a multitude of domains, tasks and languages relies on costly and time-consuming data annotation for different domain-task-language configurations.
no code implementations • CL (ACL) 2020 • Ivan Vulić, Simon Baker, Edoardo Maria Ponti, Ulla Petti, Ira Leviant, Kelly Wing, Olga Majewska, Eden Bar, Matt Malone, Thierry Poibeau, Roi Reichart, Anna Korhonen
We introduce Multi-SimLex, a large-scale lexical resource and evaluation benchmark covering data sets for 12 typologically diverse languages, including major languages (e. g., Mandarin Chinese, Spanish, Russian) as well as less-resourced ones (e. g., Welsh, Kiswahili).
no code implementations • CL (ACL) 2021 • Olga Majewska, Diana McCarthy, Jasper J. F. van den Bosch, Nikolaus Kriegeskorte, Ivan Vulić, Anna Korhonen
We demonstrate how the resultant data set can be used for fine-grained analyses and evaluation of representation learning models on the intrinsic tasks of semantic clustering and semantic similarity.
no code implementations • EMNLP 2020 • Qianchu Liu, Diana McCarthy, Anna Korhonen
One of the most powerful features of contextualized models is their dynamic embeddings for words in context, leading to state-of-the-art representations for context-aware lexical semantics.
1 code implementation • 18 Dec 2024 • Beiduo Chen, Siyao Peng, Anna Korhonen, Barbara Plank
Disagreement in human labeling is ubiquitous, and can be captured in human judgment distributions (HJDs).
no code implementations • 28 Nov 2024 • Marion Thaler, Abdullatif Köksal, Alina Leidinger, Anna Korhonen, Hinrich Schütze
Our findings reveal that biases present in pre-training data are amplified in model outputs.
1 code implementation • 19 Sep 2024 • Abdullatif Köksal, Marion Thaler, Ayyoob Imani, Ahmet Üstün, Anna Korhonen, Hinrich Schütze
Instruction tuning enhances large language models (LLMs) by aligning them with human preferences across diverse tasks.
1 code implementation • 30 Aug 2024 • Raoyuan Zhao, Abdullatif Köksal, Yihong Liu, Leonie Weissweiler, Anna Korhonen, Hinrich Schütze
In this work, we propose SYNTHEVAL, a hybrid behavioral testing framework that leverages large language models (LLMs) to generate a wide range of test types for a comprehensive evaluation of NLP models.
no code implementations • 7 Aug 2024 • Panagiotis Fytas, Anna Breger, Ian Selby, Simon Baker, Shahab Shahipasand, Anna Korhonen
Developing imaging models capable of detecting pathologies from chest X-rays can be cost and time-prohibitive for large datasets as it requires supervision to attain state-of-the-art performance.
1 code implementation • 17 Jul 2024 • Arda Yüksel, Abdullatif Köksal, Lütfi Kerem Şenel, Anna Korhonen, Hinrich Schütze
These questions are written by curriculum experts, suitable for the high-school curricula in Turkey, covering subjects ranging from natural sciences and math questions to more culturally representative topics such as Turkish Literature and the history of the Turkish Republic.
1 code implementation • 25 Jun 2024 • Beiduo Chen, Xinpeng Wang, Siyao Peng, Robert Litschko, Anna Korhonen, Barbara Plank
This study proposes to exploit LLMs to approximate HJDs using a small number of expert labels and explanations.
2 code implementations • 17 Jun 2024 • Han Zhou, Xingchen Wan, Yinhong Liu, Nigel Collier, Ivan Vulić, Anna Korhonen
Motivated by this phenomenon, we propose an automatic Zero-shot Evaluation-oriented Prompt Optimization framework, ZEPO, which aims to produce fairer preference decisions and improve the alignment of LLM evaluators with human judgments.
no code implementations • 6 Jun 2024 • Chen Cecilia Liu, Iryna Gurevych, Anna Korhonen
The surge of interest in culturally aware and adapted Natural Language Processing (NLP) has inspired much recent research.
1 code implementation • 4 Jun 2024 • Chengzu Li, Caiqi Zhang, Han Zhou, Nigel Collier, Anna Korhonen, Ivan Vulić
In this work, we thus study their capability to understand and reason over spatial relations from the top view.
1 code implementation • 15 May 2024 • Yifu Qiu, Zheng Zhao, Yftah Ziser, Anna Korhonen, Edoardo M. Ponti, Shay B. Cohen
Large language models (LLMs) often exhibit undesirable behaviours, such as generating untruthful or biased content.
no code implementations • 3 May 2024 • Yaoyiran Li, Xiang Zhai, Moustafa Alzantot, Keyi Yu, Ivan Vulić, Anna Korhonen, Mohamed Hammad
Building upon the success of Large Language Models (LLMs) in a variety of tasks, researchers have recently explored using LLMs that are pretrained on vast corpora of text for sequential recommendation.
1 code implementation • 25 Mar 2024 • Yinhong Liu, Han Zhou, Zhijiang Guo, Ehsan Shareghi, Ivan Vulić, Anna Korhonen, Nigel Collier
Large Language Models (LLMs) have demonstrated promising capabilities as automatic evaluators in assessing the quality of generated natural language.
no code implementations • 4 Mar 2024 • Evgeniia Razumovskaia, Ivan Vulić, Anna Korhonen
Supervised fine-tuning (SFT), supervised instruction tuning (SIT) and in-context learning (ICL) are three alternative, de facto standard approaches to few-shot learning.
1 code implementation • 15 Feb 2024 • Yaoyiran Li, Anna Korhonen, Ivan Vulić
Recent work has shown that, while large language models (LLMs) demonstrate strong word translation or bilingual lexicon induction (BLI) capabilities in few-shot setups, they still cannot match the performance of 'traditional' mapping-based approaches in the unsupervised scenario where no seed translation pairs are available, especially for lower-resource languages.
Bilingual Lexicon Induction Cross-Lingual Word Embeddings +9
2 code implementations • 29 Jan 2024 • Alan Ansell, Ivan Vulić, Hannah Sterz, Anna Korhonen, Edoardo M. Ponti
We experiment with instruction-tuning of LLMs on standard dataset mixtures, finding that SpIEL is often superior to popular parameter-efficient fine-tuning methods like LoRA (low-rank adaptation) in terms of performance and comparable in terms of run time.
2 code implementations • 4 Jan 2024 • Songbo Hu, Xiaobin Wang, Zhangdie Yuan, Anna Korhonen, Ivan Vulić
We present DIALIGHT, a toolkit for developing and evaluating multilingual Task-Oriented Dialogue (ToD) systems which facilitates systematic evaluations and comparisons between ToD systems using fine-tuning of Pretrained Language Models (PLMs) and those utilising the zero-shot and in-context learning capabilities of Large Language Models (LLMs).
1 code implementation • 21 Dec 2023 • Chengzu Li, Han Zhou, Goran Glavaš, Anna Korhonen, Ivan Vulić
Following the standard supervised fine-tuning (SFT) paradigm, in-context learning (ICL) has become an efficient approach propelled by the recent advancements in large language models (LLMs), yielding promising performance across various tasks in few-shot data setups.
1 code implementation • 16 Nov 2023 • Evgeniia Razumovskaia, Goran Glavaš, Anna Korhonen, Ivan Vulić
Task-oriented dialogue (ToD) systems help users execute well-defined tasks across a variety of domains (e. g., $\textit{flight booking}$ or $\textit{food ordering}$), with their Natural Language Understanding (NLU) components being dedicated to the analysis of user utterances, predicting users' intents ($\textit{Intent Detection}$, ID) and extracting values for informational slots ($\textit{Value Extraction}$, VE).
1 code implementation • 14 Nov 2023 • Yifu Qiu, Zheng Zhao, Yftah Ziser, Anna Korhonen, Edoardo M. Ponti, Shay B. Cohen
Instead, we provide LLMs with textual narratives and probe them with respect to their common-sense knowledge of the structure and duration of events, their ability to order events along a timeline, and self-consistency within their temporal model (e. g., temporal relations such as after and before are mutually exclusive for any pair of events).
no code implementations • 23 Oct 2023 • Anjali Kantharuban, Ivan Vulić, Anna Korhonen
Historically, researchers and consumers have noticed a decrease in quality when applying NLP tools to minority variants of languages (i. e. Puerto Rican Spanish or Swiss German), but studies exploring this have been limited to a select few languages.
1 code implementation • 21 Oct 2023 • Yaoyiran Li, Anna Korhonen, Ivan Vulić
Bilingual Lexicon Induction (BLI) is a core task in multilingual NLP that still, to a large extent, relies on calculating cross-lingual word representations.
Bilingual Lexicon Induction Cross-Lingual Word Embeddings +8
no code implementations • 19 Oct 2023 • Songbo Hu, Han Zhou, Moy Yuan, Milan Gritta, Guchun Zhang, Ignacio Iacobacci, Anna Korhonen, Ivan Vulić
Achieving robust language technologies that can perform well across the world's many languages is a central goal of multilingual NLP.
1 code implementation • 19 Oct 2023 • Han Zhou, Xingchen Wan, Ivan Vulić, Anna Korhonen
Prompt-based learning has been an effective paradigm for large pretrained language models (LLM), enabling few-shot or even zero-shot learning.
1 code implementation • 3 Oct 2023 • Mor Ventura, Eyal Ben-David, Anna Korhonen, Roi Reichart
Text-To-Image (TTI) models, such as DALL-E and StableDiffusion, have demonstrated remarkable prompt-based image generation capabilities.
1 code implementation • 26 Jul 2023 • Songbo Hu, Han Zhou, Mete Hergul, Milan Gritta, Guchun Zhang, Ignacio Iacobacci, Ivan Vulić, Anna Korhonen
Creating high-quality annotated data for task-oriented dialog (ToD) is known to be notoriously difficult, and the challenges are amplified when the goal is to create equitable, culturally adapted, and large-scale ToD datasets for multiple languages.
no code implementations • 5 Jun 2023 • Marinela Parović, Alan Ansell, Ivan Vulić, Anna Korhonen
We address this mismatch by exposing the task adapter to the target language adapter during training, and empirically validate several variants of the idea: in the simplest form, we alternate between using the source and target language adapters during task adapter training, which can be generalized to cycling over any set of language adapters.
1 code implementation • 2 Jun 2023 • Alan Ansell, Edoardo Maria Ponti, Anna Korhonen, Ivan Vulić
Specifically, we use a two-phase distillation approach, termed BiStil: (i) the first phase distils a general bilingual model from the MMT, while (ii) the second, task-specific phase sparsely fine-tunes the bilingual "student" model using a task-tuned variant of the original MMT as its "teacher".
no code implementations • 30 May 2023 • Yaoyiran Li, Ching-Yun Chang, Stephen Rawls, Ivan Vulić, Anna Korhonen
Research on text-to-image generation (TTI) still predominantly focuses on the English language due to the lack of annotated image-caption data in other languages; in the long run, this might widen inequitable access to TTI technology.
Cross-lingual Text-to-Image Generation Crosslingual Text-to-Image Generation +6
1 code implementation • 23 May 2023 • Yifu Qiu, Yftah Ziser, Anna Korhonen, Edoardo M. Ponti, Shay B. Cohen
With the existing faithful metrics focusing on English, even measuring the extent of this phenomenon in cross-lingual settings is hard.
1 code implementation • 22 May 2023 • Abdullatif Köksal, Omer Faruk Yalcin, Ahmet Akbiyik, M. Tahir Kilavuz, Anna Korhonen, Hinrich Schütze
For nationality as a case study, we show that LABDet `surfaces' nationality bias by training a classifier on top of a frozen PLM on non-nationality sentiment detection.
no code implementations • 22 May 2023 • Evgeniia Razumovskaia, Ivan Vulić, Anna Korhonen
It is especially effective for the most challenging transfer-free few-shot setups, paving the way for quick and data-efficient bootstrapping of multilingual slot labelers for ToD.
2 code implementations • 17 Apr 2023 • Abdullatif Köksal, Timo Schick, Anna Korhonen, Hinrich Schütze
We generate instructions via LLMs for human-written corpus examples using reverse instructions.
1 code implementation • 28 Jan 2023 • Han Zhou, Xingchen Wan, Ivan Vulić, Anna Korhonen
Large pretrained language models are widely used in downstream NLP tasks via task-specific fine-tuning, but such procedures can be costly.
no code implementations • 20 Dec 2022 • Nikita Moghe, Evgeniia Razumovskaia, Liane Guillou, Ivan Vulić, Anna Korhonen, Alexandra Birch
We use MULTI3NLU++ to benchmark state-of-the-art multilingual models for the NLU tasks of intent detection and slot labelling for TOD systems in the multilingual setting.
1 code implementation • 7 Nov 2022 • Songbo Hu, Ivan Vulić, Fangyu Liu, Anna Korhonen
At training, the high-scoring partition comprises all generated responses whose similarity to the gold response is higher than the similarity of the greedy response to the gold response.
1 code implementation • 30 Oct 2022 • Yaoyiran Li, Fangyu Liu, Ivan Vulić, Anna Korhonen
This crucial step is done via 1) creating a word similarity dataset, comprising positive word pairs (i. e., true translations) and hard negative pairs induced from the original CLWE space, and then 2) fine-tuning an mPLM (e. g., mBERT or XLM-R) in a cross-encoder manner to predict the similarity scores.
Bilingual Lexicon Induction Cross-Lingual Word Embeddings +7
1 code implementation • 12 Oct 2022 • Zhangdie Yuan, Songbo Hu, Ivan Vulić, Anna Korhonen, Zaiqiao Meng
Acquiring factual knowledge with Pretrained Language Models (PLMs) has attracted increasing attention, showing promising performance in many knowledge-intensive tasks.
no code implementations • 30 Apr 2022 • Ivan Vulić, Goran Glavaš, Fangyu Liu, Nigel Collier, Edoardo Maria Ponti, Anna Korhonen
In this work, we probe SEs for the amount of cross-lingual lexical knowledge stored in their parameters, and compare them against the original multilingual LMs.
1 code implementation • ACL 2022 • Yaoyiran Li, Fangyu Liu, Nigel Collier, Anna Korhonen, Ivan Vulić
At Stage C1, we propose to refine standard cross-lingual linear maps between static word embeddings (WEs) via a contrastive learning objective; we also show how to integrate it into the self-learning procedure for even more refined cross-lingual maps.
1 code implementation • 15 Feb 2022 • Chen Liu, Jonas Pfeiffer, Anna Korhonen, Ivan Vulić, Iryna Gurevych
2) We analyze cross-lingual VQA across different question types of varying complexity for different multilingual multimodal Transformers, and identify question types that are the most difficult to improve on.
no code implementations • 31 Jan 2022 • Olga Majewska, Evgeniia Razumovskaia, Edoardo Maria Ponti, Ivan Vulić, Anna Korhonen
Through this process we annotate a new large-scale dataset for training and evaluation of multilingual and cross-lingual ToD systems.
no code implementations • 13 Dec 2021 • Qianchu Liu, Diana McCarthy, Anna Korhonen
Our findings demonstrate that models are usually not being tested for word-in-context semantics in the same way as humans are in these tasks, which helps us better understand the model-human gap.
2 code implementations • ACL 2022 • Alan Ansell, Edoardo Maria Ponti, Anna Korhonen, Ivan Vulić
Both these masks can then be composed with the pretrained model.
1 code implementation • CoNLL (EMNLP) 2021 • Qianchu Liu, Fangyu Liu, Nigel Collier, Anna Korhonen, Ivan Vulić
Recent work indicated that pretrained language models (PLMs) such as BERT and RoBERTa can be transformed into effective sentence and word encoders even via simple self-supervised techniques.
no code implementations • IJCNLP 2019 • Edoardo Maria Ponti, Ivan Vulić, Ryan Cotterell, Roi Reichart, Anna Korhonen
Motivated by this question, we aim at constructing an informative prior over neural weights, in order to adapt quickly to held-out languages in the task of character-level language modeling.
no code implementations • ACL 2021 • Ivan Vuli{\'c}, Edoardo Maria Ponti, Anna Korhonen, Goran Glava{\v{s}}
Inspired by prior work on semantic specialization of static word embedding (WE) models, we show that it is possible to expose and enrich lexical knowledge from the LMs, that is, to specialize them to serve as effective and universal {``}decontextualized{''} word encoders even when fed input words {``}in isolation{''} (i. e., without any context).
1 code implementation • ACL 2021 • Fangyu Liu, Ivan Vulić, Anna Korhonen, Nigel Collier
To this end, we propose and evaluate a series of cross-lingual transfer methods for the XL-BEL task, and demonstrate that general-domain bitext helps propagate the available English knowledge to languages with little to no in-domain data.
no code implementations • 17 Apr 2021 • Evgeniia Razumovskaia, Goran Glavaš, Olga Majewska, Edoardo M. Ponti, Anna Korhonen, Ivan Vulić
We find that the most critical factor preventing the creation of truly multilingual ToD systems is the lack of datasets in most languages for both training and evaluation.
1 code implementation • EMNLP 2021 • Qianchu Liu, Edoardo M. Ponti, Diana McCarthy, Ivan Vulić, Anna Korhonen
In order to address these gaps, we present AM2iCo (Adversarial and Multilingual Meaning in Context), a wide-coverage cross-lingual and multilingual evaluation set; it aims to faithfully assess the ability of state-of-the-art (SotA) representation models to understand the identity of word meaning in cross-lingual contexts for 14 language pairs.
1 code implementation • EMNLP 2021 • Fangyu Liu, Ivan Vulić, Anna Korhonen, Nigel Collier
In this work, we demonstrate that it is possible to turn MLMs into effective universal lexical and sentence encoders even without any additional data and without any supervision.
Ranked #15 on Semantic Textual Similarity on STS16
Contrastive Learning Cross-Lingual Semantic Textual Similarity +5
1 code implementation • EACL 2021 • Yi Zhu, Ehsan Shareghi, Yingzhen Li, Roi Reichart, Anna Korhonen
Semi-supervised learning through deep generative models and multi-lingual pretraining techniques have orchestrated tremendous success across different areas of NLP.
no code implementations • ACL 2021 • Olga Majewska, Ivan Vulić, Goran Glavaš, Edoardo M. Ponti, Anna Korhonen
We investigate whether injecting explicit information on verbs' semantic-syntactic behaviour improves the performance of LM-pretrained Transformers in event extraction tasks -- downstream tasks for which accurate verb processing is paramount.
no code implementations • ACL 2021 • Mengjie Zhao, Yi Zhu, Ehsan Shareghi, Ivan Vulić, Roi Reichart, Anna Korhonen, Hinrich Schütze
Few-shot crosslingual transfer has been shown to outperform its zero-shot counterpart with pretrained encoders like multilingual BERT.
1 code implementation • COLING 2020 • Olga Majewska, Ivan Vuli{\'c}, Diana McCarthy, Anna Korhonen
We present the first evaluation of the applicability of a spatial arrangement method (SpAM) to a typologically diverse language sample, and its potential to produce semantic evaluation resources to support multilingual NLP, with a focus on verb semantics.
no code implementations • SEMEVAL 2020 • Goran Glava{\v{s}}, Ivan Vuli{\'c}, Anna Korhonen, Simone Paolo Ponzetto
The shared task spans three dimensions: (1) monolingual vs. cross-lingual LE, (2) binary vs. graded LE, and (3) a set of 6 diverse languages (and 15 corresponding language pairs).
1 code implementation • COLING 2020 • Yaoyiran Li, Edoardo M. Ponti, Ivan Vulić, Anna Korhonen
On the other hand, this also provides an extrinsic evaluation protocol to probe the properties of emergent languages ex vitro.
no code implementations • EMNLP 2020 • Ivan Vulić, Edoardo Maria Ponti, Robert Litschko, Goran Glavaš, Anna Korhonen
The success of large pretrained language models (LMs) such as BERT and RoBERTa has sparked interest in probing their representations, in order to unveil what types of knowledge they implicitly capture.
no code implementations • ACL 2020 • Mladen Karan, Ivan Vuli{\'c}, Anna Korhonen, Goran Glava{\v{s}}
Effective projection-based cross-lingual word embedding (CLWE) induction critically relies on the iterative self-learning procedure.
no code implementations • ACL 2020 • Ryohei Sasano, Anna Korhonen
This paper presents an investigation on the distribution of word vectors belonging to a certain word class in a pre-trained word vector space.
no code implementations • WS 2020 • Ivan Vuli{\'c}, Anna Korhonen, Goran Glava{\v{s}}
Work on projection-based induction of cross-lingual word embedding spaces (CLWEs) predominantly focuses on the improvement of the projection (i. e., mapping) mechanisms.
1 code implementation • ACL 2020 • Daniela Gerz, Ivan Vulić, Marek Rei, Roi Reichart, Anna Korhonen
We present a neural framework for learning associations between interrelated groups of words such as the ones found in Subject-Verb-Object (SVO) structures.
1 code implementation • EMNLP 2020 • Edoardo Maria Ponti, Goran Glavaš, Olga Majewska, Qianchu Liu, Ivan Vulić, Anna Korhonen
In order to simulate human language capacity, natural language processing systems must be able to reason about the dynamics of everyday situations, including their possible causes and effects.
Ranked #3 on Cross-Lingual Transfer on XCOPA (using extra training data)
no code implementations • LREC 2020 • Olga Majewska, Diana McCarthy, Jasper van den Bosch, Nikolaus Kriegeskorte, Ivan Vuli{\'c}, Anna Korhonen
We present a novel methodology for fast bottom-up creation of large-scale semantic similarity resources to support development and evaluation of NLP systems.
no code implementations • 5 Apr 2020 • Yixuan Su, Deng Cai, Yan Wang, Simon Baker, Anna Korhonen, Nigel Collier, Xiaojiang Liu
To enable better balance between the content quality and the style, we introduce a new training strategy, know as Information-Guided Reinforcement Learning (IG-RL).
no code implementations • 5 Apr 2020 • Yixuan Su, Yan Wang, Simon Baker, Deng Cai, Xiaojiang Liu, Anna Korhonen, Nigel Collier
A stylistic response generator then takes the prototype and the desired language style as model input to obtain a high-quality and stylistic response.
no code implementations • 10 Mar 2020 • Ivan Vulić, Simon Baker, Edoardo Maria Ponti, Ulla Petti, Ira Leviant, Kelly Wing, Olga Majewska, Eden Bar, Matt Malone, Thierry Poibeau, Roi Reichart, Anna Korhonen
We introduce Multi-SimLex, a large-scale lexical resource and evaluation benchmark covering datasets for 12 typologically diverse languages, including major languages (e. g., Mandarin Chinese, Spanish, Russian) as well as less-resourced ones (e. g., Welsh, Kiswahili).
no code implementations • EMNLP 2020 • Haim Dubossarsky, Ivan Vulić, Roi Reichart, Anna Korhonen
Performance in cross-lingual NLP tasks is impacted by the (dis)similarity of languages at hand: e. g., previous work has suggested there is a connection between the expected success of bilingual lexicon induction (BLI) and the assumption of (approximate) isomorphism between monolingual embedding spaces.
1 code implementation • 30 Jan 2020 • Edoardo M. Ponti, Ivan Vulić, Ryan Cotterell, Marinela Parovic, Roi Reichart, Anna Korhonen
In this work, we propose a Bayesian generative model for the space of neural parameters.
no code implementations • 26 Nov 2019 • Bo-Hsiang Tseng, Marek Rei, Paweł Budzianowski, Richard E. Turner, Bill Byrne, Anna Korhonen
Dialogue systems benefit greatly from optimizing on detailed annotations, such as transcribed utterances, internal dialogue state representations and dialogue act labels.
no code implementations • CONLL 2019 • Qianchu Liu, Diana McCarthy, Ivan Vuli{\'c}, Anna Korhonen
In this paper, we present a thorough investigation on methods that align pre-trained contextualized embeddings into shared cross-lingual context-aware embedding space, providing strong reference benchmarks for future context-aware crosslingual models.
no code implementations • IJCNLP 2019 • Edoardo Maria Ponti, Ivan Vuli{\'c}, Goran Glava{\v{s}}, Roi Reichart, Anna Korhonen
Semantic specialization integrates structured linguistic knowledge from external resources (such as lexical relations in WordNet) into pretrained distributional vectors in the form of constraints.
no code implementations • IJCNLP 2019 • Bo-Hsiang Tseng, Marek Rei, Pawe{\l} Budzianowski, Richard Turner, Bill Byrne, Anna Korhonen
Dialogue systems benefit greatly from optimizing on detailed annotations, such as transcribed utterances, internal dialogue state representations and dialogue act labels.
no code implementations • CONLL 2019 • Yi Zhu, Benjamin Heinzerling, Ivan Vulić, Michael Strube, Roi Reichart, Anna Korhonen
Recent work has validated the importance of subword information for word representation learning.
1 code implementation • COLING 2020 • Anne Lauscher, Ivan Vulić, Edoardo Maria Ponti, Anna Korhonen, Goran Glavaš
In this work, we complement such distributional knowledge with external lexical knowledge, that is, we integrate the discrete knowledge on word-level semantic similarity into pretraining.
1 code implementation • IJCNLP 2019 • Ivan Vulić, Goran Glavaš, Roi Reichart, Anna Korhonen
A series of bilingual lexicon induction (BLI) experiments with 15 diverse languages (210 language pairs) show that fully unsupervised CLWE methods still fail for a large number of language pairs (e. g., they yield zero BLI performance for 87/210 pairs).
1 code implementation • WS 2019 • Billy Chiu, Simon Baker, Martha Palmer, Anna Korhonen
Verbs play a fundamental role in many biomed-ical tasks and applications such as relation and event extraction.
no code implementations • NAACL 2019 • Ehsan Shareghi, Yingzhen Li, Yi Zhu, Roi Reichart, Anna Korhonen
While neural dependency parsers provide state-of-the-art accuracy for several languages, they still rely on large amounts of costly labeled training data.
no code implementations • NAACL 2019 • Ehsan Shareghi, Daniela Gerz, Ivan Vuli{\'c}, Anna Korhonen
In recent years neural language models (LMs) have set the state-of-the-art performance for several benchmarking datasets.
no code implementations • SEMEVAL 2019 • Qianchu Liu, Diana McCarthy, Anna Korhonen
There is a growing awareness of the need to handle rare and unseen words in word representation modelling.
1 code implementation • NAACL 2019 • Yi Zhu, Ivan Vulić, Anna Korhonen
The use of subword-level information (e. g., characters, character n-grams, morphemes) has become ubiquitous in modern word representation learning.
no code implementations • EMNLP 2018 • Daniela Gerz, Ivan Vuli{\'c}, Edoardo Maria Ponti, Roi Reichart, Anna Korhonen
A key challenge in cross-lingual NLP is developing general language-independent architectures that are equally applicable to any language.
1 code implementation • EMNLP 2018 • Edoardo Maria Ponti, Ivan Vulić, Goran Glavaš, Nikola Mrkšić, Anna Korhonen
Our adversarial post-specialization method propagates the external lexical knowledge to the full distributional space.
no code implementations • CL 2019 • Edoardo Maria Ponti, Helen O'Horan, Yevgeni Berzak, Ivan Vulić, Roi Reichart, Thierry Poibeau, Ekaterina Shutova, Anna Korhonen
Linguistic typology aims to capture structural and semantic variation across the world's languages.
no code implementations • ACL 2018 • Edoardo Maria Ponti, Roi Reichart, Anna Korhonen, Ivan Vuli{\'c}
The transfer or share of knowledge between languages is a potential solution to resource scarcity in NLP.
1 code implementation • NAACL 2018 • Ivan Vulić, Goran Glavaš, Nikola Mrkšić, Anna Korhonen
Word vector specialisation (also known as retrofitting) is a portable, light-weight approach to fine-tuning arbitrary distributional word vector spaces by injecting external knowledge from rich lexical resources such as WordNet.
no code implementations • TACL 2018 • Daniela Gerz, Ivan Vuli{\'c}, Edoardo Ponti, Jason Naradowsky, Roi Reichart, Anna Korhonen
Neural architectures are prominent in the construction of language models (LMs).
no code implementations • 1 Nov 2017 • Yiding Lu, Yufan Guo, Anna Korhonen
Conclusion: This demonstrates that approaches purely based on network topology provide a more suitable approach to DTI prediction in the many real-life situations where little or no prior knowledge is available about the characteristics of drugs, targets, or their interactions.
1 code implementation • BMC Bioinformatics 2017 • Gamal Crichton, Sampo Pyysalo, Billy Chiu, Anna Korhonen
Additionally, we investigated the effect of dataset size on performance in both single- and multi-task settings.
no code implementations • WS 2017 • Simon Baker, Anna Korhonen
Many tasks in the biomedical domain require the assignment of one or more predefined labels to input text, where the labels are a part of a hierarchical structure (such as a taxonomy).
no code implementations • EMNLP 2017 • Ivan Vulić, Nikola Mrkšić, Anna Korhonen
Existing approaches to automatic VerbNet-style verb classification are heavily dependent on feature engineering and therefore limited to languages with mature NLP pipelines.
2 code implementations • 1 Jun 2017 • Nikola Mrkšić, Ivan Vulić, Diarmuid Ó Séaghdha, Ira Leviant, Roi Reichart, Milica Gašić, Anna Korhonen, Steve Young
We present Attract-Repel, an algorithm for improving the semantic quality of word vectors by injecting constraints extracted from lexical resources.
no code implementations • ACL 2017 • Ivan Vulić, Nikola Mrkšić, Roi Reichart, Diarmuid Ó Séaghdha, Steve Young, Anna Korhonen
Morphologically rich languages accentuate two properties of distributional vector space models: 1) the difficulty of inducing accurate representations for low-frequency word forms; and 2) insensitivity to distinct lexical relations that have similar distributional signatures.
no code implementations • SEMEVAL 2017 • Edoardo Maria Ponti, Ivan Vulić, Anna Korhonen
Distributed representations of sentences have been developed recently to represent their meaning as real-valued vectors.
no code implementations • EACL 2017 • Ivan Vuli{\'c}, Douwe Kiela, Anna Korhonen
Recent work on evaluating representation learning architectures in NLP has established a need for evaluation protocols based on subconscious cognitive measures rather than manually tailored intrinsic similarity and relatedness tasks.
no code implementations • WS 2017 • Edoardo Maria Ponti, Anna Korhonen
Causal relations play a key role in information extraction and reasoning.
no code implementations • TACL 2017 • Nikola Mrk{\v{s}}i{\'c}, Ivan Vuli{\'c}, Diarmuid {\'O} S{\'e}aghdha, Ira Leviant, Roi Reichart, Milica Ga{\v{s}}i{\'c}, Anna Korhonen, Steve Young
We present Attract-Repel, an algorithm for improving the semantic quality of word vectors by injecting constraints extracted from lexical resources.
no code implementations • COLING 2016 • Simon Baker, Douwe Kiela, Anna Korhonen
The conventional solution for handling sparsely labelled data is extensive feature engineering.
no code implementations • WS 2016 • Simon Baker, Anna Korhonen, Sampo Pyysalo
Methods based on deep learning approaches have recently achieved state-of-the-art performance in a range of machine learning tasks and are increasingly applied to natural language processing (NLP).
no code implementations • COLING 2016 • Helen O'Horan, Yevgeni Berzak, Ivan Vulić, Roi Reichart, Anna Korhonen
In recent years linguistic typology, which classifies the world's languages according to their functional and structural properties, has been widely used to support multilingual NLP.
no code implementations • CONLL 2017 • Ivan Vulić, Roy Schwartz, Ari Rappoport, Roi Reichart, Anna Korhonen
With our selected context configurations, we train on only 14% (A), 26. 2% (V), and 33. 6% (N) of all dependency-based contexts, resulting in a reduced training time.
no code implementations • CL 2017 • Ivan Vulić, Daniela Gerz, Douwe Kiela, Felix Hill, Anna Korhonen
We introduce HyperLex - a dataset and evaluation resource that quantifies the extent of of the semantic category membership, that is, type-of relation also known as hyponymy-hypernymy or lexical entailment (LE) relation between 2, 616 concept pairs.
1 code implementation • EMNLP 2016 • Daniela Gerz, Ivan Vulić, Felix Hill, Roi Reichart, Anna Korhonen
Verbs play a critical role in the meaning of sentences, but these ubiquitous words have received little attention in recent distributional semantics research.
no code implementations • EMNLP 2016 • Yevgeni Berzak, Yan Huang, Andrei Barbu, Anna Korhonen, Boris Katz
Our agreement results control for parser bias, and are consequential in that they are on par with state of the art parsing performance for English newswire.
1 code implementation • NAACL 2016 • Felix Hill, Kyunghyun Cho, Anna Korhonen
Unsupervised methods for learning distributed representations of words are ubiquitous in today's NLP research, but far less is known about the best ways to learn distributed phrase or sentence representations from unlabelled data.
Ranked #16 on Subjectivity Analysis on SUBJ
no code implementations • 9 Oct 2015 • Simon Baker, Ilona Silins, Yufan Guo, Imran Ali, Johan Högberg, Ulla Stenius, Anna Korhonen
The hallmarks of cancer have become highly influential in cancer research.
3 code implementations • TACL 2016 • Felix Hill, Kyunghyun Cho, Anna Korhonen, Yoshua Bengio
Distributional models that learn rich semantic word representations are a success story of recent NLP research.
no code implementations • TACL 2015 • Yufan Guo, Roi Reichart, Anna Korhonen
Inferring the information structure of scientific documents is useful for many NLP applications.
3 code implementations • CL 2015 • Felix Hill, Roi Reichart, Anna Korhonen
We present SimLex-999, a gold standard resource for evaluating distributional semantic models that improves on existing resources in several important ways.
no code implementations • LREC 2014 • Xiao Jiang, Yufan Guo, Jeroen Geertzen, Dora Alexopoulou, Lin Sun, Anna Korhonen
Native Language Identification (NLI) is a task aimed at determining the native language (L1) of learners of second language (L2) on the basis of their written texts.
BIG-bench Machine Learning Native Language Identification +1
no code implementations • TACL 2014 • Felix Hill, Roi Reichart, Anna Korhonen
Multi-modal models that learn semantic representations from both linguistic and perceptual input outperform language-only models on a range of evaluations, and better reflect human concept acquisition.