1 code implementation • EMNLP 2021 • Asahi Ushio, Jose Camacho-Collados, Steven Schockaert
Among others, this makes it possible to distill high-quality word vectors from pre-trained language models.
no code implementations • 23 Oct 2023 • Dimosthenis Antypas, Asahi Ushio, Francesco Barbieri, Leonardo Neves, Kiamehr Rezaee, Luis Espinosa-Anke, Jiaxin Pei, Jose Camacho-Collados
Despite its relevance, the maturity of NLP for social media pales in comparison with general-purpose models, metrics and benchmarks.
1 code implementation • 30 Sep 2023 • Asahi Ushio, Jose Camacho-Collados, Steven Schockaert
In particular, we show that masked language models such as RoBERTa can be straightforwardly fine-tuned for this purpose, using only a small amount of training data.
1 code implementation • 27 May 2023 • Asahi Ushio, Fernando Alva-Manchego, Jose Camacho-Collados
Generating questions along with associated answers from a text has applications in several domains, such as creating reading comprehension tests for students, or improving document search by providing auxiliary questions and answers based on the query.
1 code implementation • 26 May 2023 • Asahi Ushio, Fernando Alva-Manchego, Jose Camacho-Collados
This task has a variety of applications, such as data augmentation for question answering (QA) models, information retrieval and education.
1 code implementation • 24 May 2023 • Asahi Ushio, Yi Zhou, Jose Camacho-Collados
Multilingual language model (LM) have become a powerful tool in NLP especially for non-English languages.
no code implementations • 24 May 2023 • Asahi Ushio, Jose Camacho Collados, Steven Schockaert
Relations such as "is influenced by", "is known for" or "is a competitor of" are inherently graded: we can rank entity pairs based on how well they satisfy these relations, but it is hard to draw a line between those pairs that satisfy them and those that do not.
1 code implementation • 8 Oct 2022 • Asahi Ushio, Fernando Alva-Manchego, Jose Camacho-Collados
It includes general-purpose datasets such as SQuAD for English, datasets from ten domains and two styles, as well as datasets in eight different languages.
1 code implementation • 7 Oct 2022 • Asahi Ushio, Leonardo Neves, Vitor Silva, Francesco Barbieri, Jose Camacho-Collados
Recent progress in language model pre-training has led to important improvements in Named Entity Recognition (NER).
no code implementations • COLING 2022 • Dimosthenis Antypas, Asahi Ushio, Jose Camacho-Collados, Leonardo Neves, Vítor Silva, Francesco Barbieri
Social media platforms host discussions about a wide variety of topics that arise everyday.
1 code implementation • EACL 2021 • Asahi Ushio, Jose Camacho-Collados
In this paper, we present T-NER (Transformer-based Named Entity Recognition), a Python library for NER LM finetuning.
Ranked #4 on Named Entity Recognition (NER) on WNUT 2017
1 code implementation • 29 Jun 2022 • Jose Camacho-Collados, Kiamehr Rezaee, Talayeh Riahi, Asahi Ushio, Daniel Loureiro, Dimosthenis Antypas, Joanne Boisson, Luis Espinosa-Anke, Fangyu Liu, Eugenio Martínez-Cámara, Gonzalo Medina, Thomas Buhrmann, Leonardo Neves, Francesco Barbieri
In this paper we present TweetNLP, an integrated platform for Natural Language Processing (NLP) in social media.
no code implementations • 17 Nov 2021 • Aleksandra Edwards, Asahi Ushio, Jose Camacho-Collados, Hélène de Ribaupierre, Alun Preece
Data augmentation techniques are widely used for enhancing the performance of machine learning models by tackling class imbalance issues and data sparsity.
1 code implementation • 21 Sep 2021 • Asahi Ushio, Jose Camacho-Collados, Steven Schockaert
Among others, this makes it possible to distill high-quality word vectors from pre-trained language models.
1 code implementation • ACL 2021 • Asahi Ushio, Luis Espinosa-Anke, Steven Schockaert, Jose Camacho-Collados
Analogies play a central role in human commonsense reasoning.
1 code implementation • EMNLP 2021 • Asahi Ushio, Federico Liberatore, Jose Camacho-Collados
Term weighting schemes are widely used in Natural Language Processing and Information Retrieval.
no code implementations • 6 Feb 2019 • Max F. Frenzel, Bogdan Teleaga, Asahi Ushio
Deep generative models are universal tools for learning data distributions on high dimensional data spaces via a mapping to lower dimensional latent spaces.