1 code implementation • NLPerspectives (LREC) 2022 • Laura Biester, Vanita Sharma, Ashkan Kazemi, Naihao Deng, Steven Wilson, Rada Mihalcea
Recent studies have shown that for subjective annotation tasks, the demographics, lived experiences, and identity of annotators can have a large impact on how items are labeled.
1 code implementation • SemEval (NAACL) 2022 • Ibrahim Abu Farha, Silviu Vlad Oprea, Steven Wilson, Walid Magdy
Most of the participating teams utilised pre-trained language models.
no code implementations • NAACL (WNU) 2022 • Achyutarama Ganti, Steven Wilson, Zexin Ma, Xinyan Zhao, Rong Ma
Narratives have been shown to be an effective way to communicate health risks and promote health behavior change, and given the growing amount of health information being shared on social media, it is crucial to study health-related narratives in social media.
no code implementations • ACL 2022 • Silviu Vlad Oprea, Steven Wilson, Walid Magdy
Previous sarcasm generation research has focused on how to generate text that people perceive as sarcastic to create more human-like interactions.
no code implementations • EMNLP (NLP+CSS) 2020 • Jinhang Li, Giorgos Longinos, Steven Wilson, Walid Magdy
Emoji are widely used to express emotions and concepts on social media, and prior work has shown that users’ choice of emoji reflects the way that they wish to present themselves to the world.
no code implementations • EMNLP (NLP+CSS) 2020 • Felix Hennig, Steven Wilson
Previous English-language diachronic change models based on word embeddings have typically used single tokens to represent entities, including names of people.
no code implementations • EMNLP (ACL) 2021 • Silviu Oprea, Steven Wilson, Walid Magdy
We introduce Chandler, a system that generates sarcastic responses to a given utterance.
no code implementations • EMNLP (insights) 2020 • Steven Wilson, Walid Magdy, Barbara McGillivray, Gareth Tyson
Previous work has shown how to effectively use external resources such as dictionaries to improve English-language word embeddings, either by manipulating the training process or by applying post-hoc adjustments to the embedding space.
no code implementations • SEMEVAL 2021 • J. A. Meaney, Steven Wilson, Luis Chiruzzo, Adam Lopez, Walid Magdy
Our subtasks were binary humor detection, prediction of humor and offense ratings, and a novel controversy task: to predict if the variance in the humor ratings was higher than a specific threshold.
no code implementations • SEMEVAL 2020 • J. A. Meaney, Steven Wilson, Walid Magdy
The use of pre-trained language models such as BERT and ULMFiT has become increasingly popular in shared tasks, due to their powerful language modelling capabilities.
no code implementations • LREC 2020 • Amy Rechkemmer, Steven Wilson, Rada Mihalcea
Using a set of over 2 million posts from distinct Twitter users around the country dating back as far as 2014, we ask the following question: is there a difference in how Americans express themselves online depending on whether they reside in an urban or rural area?
no code implementations • LREC 2020 • Steven Wilson, Walid Magdy, Barbara McGillivray, Kiran Garimella, Gareth Tyson
The choice of the corpus on which word embeddings are trained can have a sizable effect on the learned representations, the types of analyses that can be performed with them, and their utility as features for machine learning models.
no code implementations • IJCNLP 2017 • Steven Wilson, Rada Mihalcea
The things people do in their daily lives can provide valuable insights into their personality, values, and interests.
no code implementations • 20 Dec 2016 • Konstantinos Pappas, Steven Wilson, Rada Mihalcea
People's personality and motivations are manifest in their everyday language usage.