no code implementations • COLING (PEOPLES) 2020 • Dirk Johannßen, Chris Biemann
We employ this model to investigate a change of language towards social unrest during the COVID-19 pandemic by comparing established psychological predictors on samples of tweets from spring 2019 with spring 2020.
no code implementations • RaPID (LREC) 2022 • Dirk Johannßen, Chris Biemann, David Scheffer
It can be assumed that the Jungian psychology types of extraverts and introverts react differently to these challenges.
1 code implementation • SIGUL (LREC) 2022 • Tadesse Destaw, Seid Muhie Yimam, Abinew Ayele, Chris Biemann
Questions are posted in Amharic, English, or Amharic but in a Latin script.
1 code implementation • NoDaLiDa 2021 • Timo Johner, Abhik Jana, Chris Biemann
Recent research using pre-trained language models for multi-document summarization task lacks deep investigation of potential erroneous cases and their possible application on other languages.
no code implementations • LREC 2022 • Meriem Beloucif, Seid Muhie Yimam, Steffen Stahlhacke, Chris Biemann
Comparative Question Answering (cQA) is the task of providing concrete and accurate responses to queries such as: “Is Lyft cheaper than a regular taxi?” or “What makes a mortgage different from a regular loan?”.
no code implementations • LREC 2022 • Xintong Wang, Florian Schneider, Özge Alacam, Prateek Chaudhury, Chris Biemann
In this respect, several traditional NLP approaches can assist L2 readers in text comprehension processes, such as simplifying text or giving dictionary descriptions for complex words.
no code implementations • LREC 2022 • Fynn Petersen-Frey, Marcus Soll, Louis Kobras, Melf Johannsen, Peter Kling, Chris Biemann
We present a dataset containing source code solutions to algorithmic programming exercises solved by hundreds of Bachelor-level students at the University of Hamburg.
2 code implementations • KONVENS (WS) 2022 • Robert Geislinger, Benjamin Milde, Chris Biemann
Ranked #2 on Speech Recognition on TUDA (using extra training data)
no code implementations • NAACL (PrivateNLP) 2021 • Abhik Jana, Chris Biemann
In this paper, we investigate the applicability of a privacy-preserving framework for sequence tagging tasks, specifically NER.
no code implementations • PaM 2020 • Saba Anwar, Artem Shelmanov, Alexander Panchenko, Chris Biemann
We investigate a simple yet effective method, lexical substitution with word representation models, to automatically expand a small set of frame-annotated sentences with new words for their respective roles and LUs.
1 code implementation • Findings (EMNLP) 2021 • Meriem Beloucif, Chris Biemann
Pretrained language models (PTLMs) yield state-of-the-art performance on many natural language processing tasks, including syntax, semantics and commonsense.
no code implementations • 9 Jan 2025 • Gregor Geigle, Florian Schneider, Carolin Holtermann, Chris Biemann, Radu Timofte, Anne Lauscher, Goran Glavaš
Most Large Vision-Language Models (LVLMs) to date are trained predominantly on English data, which makes them struggle to understand non-English input and fail to generate output in the desired target language.
no code implementations • 23 Oct 2024 • Xintong Wang, Jingheng Pan, Longqin Jiang, Liang Ding, Xingshan Li, Chris Biemann
Despite their impressive capabilities, large language models (LLMs) often lack interpretability and can generate toxic content.
no code implementations • 18 Oct 2024 • Thennal D K, Tim Fischer, Chris Biemann
We evaluate four different state-of-the-art LLMs on text embedding tasks and find that our method can prune up to 30\% of layers with negligible impact on performance and up to 80\% with only a modest drop.
no code implementations • 27 Jun 2024 • Seid Muhie Yimam, Daryna Dementieva, Tim Fischer, Daniil Moskovskiy, Naquee Rizwan, Punyajoy Saha, Sarthak Roy, Martin Semmann, Alexander Panchenko, Chris Biemann, Animesh Mukherjee
Despite regulations imposed by nations and social media platforms, such as recent EU regulations targeting digital violence, abusive content persists as a significant challenge.
2 code implementations • 18 Jun 2024 • Viktor Moskvoretskii, Nazarii Tupitsa, Chris Biemann, Samuel Horváth, Eduard Gorbunov, Irina Nikishina
We present a new approach called MeritOpt based on the Personalized Federated Learning algorithm MeritFed that can be applied to Natural Language Tasks with heterogeneous data.
no code implementations • 25 Apr 2024 • Fynn Petersen-Frey, Chris Biemann
We specify our annotation schema, describe the creation of the dataset and provide a quantitative analysis.
1 code implementation • 18 Apr 2024 • Abinew Ali Ayele, Esubalew Alemneh Jalew, Adem Chanie Ali, Seid Muhie Yimam, Chris Biemann
The prevalence of digital media and evolving sociopolitical dynamics have significantly amplified the dissemination of hateful content.
2 code implementations • 27 Mar 2024 • Xintong Wang, Jingheng Pan, Liang Ding, Chris Biemann
Our method is inspired by our observation that what we call disturbance instructions significantly exacerbate hallucinations in multimodal fusion modules.
1 code implementation • 22 Mar 2024 • Punyajoy Saha, Aalok Agrawal, Abhik Jana, Chris Biemann, Animesh Mukherjee
In terms of prompting, we find that our proposed strategies help in improving counter speech generation across all the models.
2 code implementations • 13 Feb 2024 • Nedjma Ousidhoum, Shamsuddeen Hassan Muhammad, Mohamed Abdalla, Idris Abdulmumin, Ibrahim Said Ahmad, Sanchit Ahuja, Alham Fikri Aji, Vladimir Araujo, Abinew Ali Ayele, Pavan Baswani, Meriem Beloucif, Chris Biemann, Sofia Bourhim, Christine de Kock, Genet Shanko Dekebo, Oumaima Hourrane, Gopichand Kanumolu, Lokesh Madasu, Samuel Rutunda, Manish Shrivastava, Thamar Solorio, Nirmal Surange, Hailegnaw Getaneh Tilaye, Krishnapriya Vishnubhotla, Genta Winata, Seid Muhie Yimam, Saif M. Mohammad
Exploring and quantifying semantic relatedness is central to representing language and holds significant implications across various NLP tasks.
no code implementations • 8 Oct 2023 • Xintong Wang, Xiaoyu Li, Xingshan Li, Chris Biemann
Large Language Models (LLMs) have emerged as dominant foundational models in modern NLP.
1 code implementation • 14 Sep 2023 • Debayan Banerjee, Arefa, Ricardo Usbeck, Chris Biemann
In this work, we present a web application named DBLPLink, which performs entity linking over the DBLP scholarly knowledge graph.
1 code implementation • 24 May 2023 • Debayan Banerjee, Pranav Ajit Nair, Ricardo Usbeck, Chris Biemann
In this work, we analyse the role of output vocabulary for text-to-text (T2T) models on the task of SPARQL semantic parsing.
1 code implementation • 23 Mar 2023 • Debayan Banerjee, Pranav Ajit Nair, Ricardo Usbeck, Chris Biemann
To further improve the results, we instruct the model to produce a truncated version of the KG embedding for each entity.
1 code implementation • 23 Mar 2023 • Debayan Banerjee, Sushil Awale, Ricardo Usbeck, Chris Biemann
In this work we create a question answering dataset over the DBLP scholarly knowledge graph (KG).
no code implementations • 28 Jan 2023 • Debayan Banerjee, Mathis Poser, Christina Wiethof, Varun Shankar Subramanian, Richard Paucar, Eva A. C. Bittner, Chris Biemann
AI enabled chat bots have recently been put to use to answer customer service queries, however it is a common feedback of users that bots lack a personal touch and are often unable to understand the real intent of the user's question.
no code implementations • 25 Jan 2023 • Debayan Banerjee, Seid Muhie Yimam, Sushil Awale, Chris Biemann
In this work, we present ARDIAS, a web-based application that aims to provide researchers with a full suite of discovery and collaboration tools.
1 code implementation • 27 Apr 2022 • Debayan Banerjee, Pranav Ajit Nair, Jivat Neet Kaur, Ricardo Usbeck, Chris Biemann
In this work, we focus on the task of generating SPARQL queries from natural language questions, which can then be executed on Knowledge Graphs (KGs).
no code implementations • EACL 2021 • Christian Haase, Saba Anwar, Seid Muhie Yimam, Alexander Friedrich, Chris Biemann
There are two main approaches to the exploration of dynamic networks: the discrete one compares a series of clustered graphs from separate points in time.
no code implementations • 2 Feb 2022 • Markus J. Hofmann, Steffen Remus, Chris Biemann, Ralph Radach, Lars Kuchinke
(3) In recurrent neural networks (RNNs), the subsymbolic units are trained to predict the next word, given all preceding words in the sentences.
1 code implementation • KONVENS (WS) 2021 • Niklas von Boguszewski, Sana Moin, Anirban Bhowmick, Seid Muhie Yimam, Chris Biemann
Hence, we show that transfer learning from the social media domain is efficacious in classifying hate and offensive speech in movies through subtitles.
no code implementations • NAACL 2021 • Max Wiechmann, Seid Muhie Yimam, Chris Biemann
ActiveAnno is built with extensible design and easy deployment in mind, all to enable users to perform annotation tasks with high efficiency and high-quality annotation results.
no code implementations • NAACL 2021 • Florian Schneider, {\"O}zge Ala{\c{c}}am, Xintong Wang, Chris Biemann
In primary school, children{'}s books, as well as in modern language learning apps, multi-modal learning strategies like illustrations of terms and phrases are used to support reading comprehension.
no code implementations • NAACL 2021 • Hans Ole Hatzel, Chris Biemann
In this thesis proposal, we explore the application of event extraction to literary texts.
no code implementations • NAACL 2021 • Sian Gooding, Ekaterina Kochmar, Seid Muhie Yimam, Chris Biemann
Lexical complexity is a highly subjective notion, yet this factor is often neglected in lexical simplification and readability systems which use a {''}one-size-fits-all{''} approach.
no code implementations • EACL 2021 • Marlo Haering, Jakob Smedegaard Andersen, Chris Biemann, Wiebke Loosen, Benjamin Milde, Tim Pietz, Christian St{\"o}cker, Gregor Wiedemann, Olaf Zukunft, Walid Maalej
With the increasing number of user comments in diverse domains, including comments on online journalism and e-commerce websites, the manual content analysis of these comments becomes time-consuming and challenging.
6 code implementations • 18 Dec 2020 • Binny Mathew, Punyajoy Saha, Seid Muhie Yimam, Chris Biemann, Pawan Goyal, Animesh Mukherjee
We also observe that models, which utilize the human rationales for training, perform better in reducing unintended bias towards target communities.
Ranked #3 on Hate Speech Detection on HateXplain
no code implementations • 8 Dec 2020 • Dirk Johannßen, Chris Biemann
We employ this model to investigate a change of language towards social unrest during the COVID-19 pandemic by comparing established psychological predictors on samples of tweets from spring 2019 with spring 2020.
no code implementations • COLING 2020 • Seid Muhie Yimam, Hizkiel Mitiku Alemayehu, Abinew Ayele, Chris Biemann
To advance the sentiment analysis research in Amharic and other related low-resource languages, we release the dataset, the annotation tool, source code, and models publicly under a permissive.
1 code implementation • 2 Nov 2020 • Seid Muhie Yimam, Abinew Ali Ayele, Gopalakrishnan Venkatesh, Chris Biemann
We find that newly trained models perform better than pre-trained multilingual models.
no code implementations • COLING (CogALex) 2020 • Markus J. Hofmann, Lara Müller, Andre Rölke, Ralph Radach, Chris Biemann
Then we trained word2vec models from individual corpora and a 70 million-sentence newspaper corpus to obtain individual and norm-based long-term memory structure.
no code implementations • ACL 2020 • Fynn Schr{\"o}der, Chris Biemann
We propose new methods to automatically assess the similarity of sequence tagging datasets to identify beneficial auxiliary data for MTL or TL setups.
no code implementations • 31 May 2020 • Ozge Sevgili, Artem Shelmanov, Mikhail Arkhipov, Alexander Panchenko, Chris Biemann
This survey presents a comprehensive description of recent neural entity linking (EL) systems developed since 2015 as a result of the "deep learning revolution" in natural language processing.
1 code implementation • 29 May 2020 • Benjamin Milde, Chris Biemann
The model is trained on 600h and 6000h of English read speech.
Ranked #2 on Speech Recognition on Libri-Light test-other (ABX-within metric)
no code implementations • SEMEVAL 2020 • Gregor Wiedemann, Seid Muhie Yimam, Chris Biemann
Fine-tuning of pre-trained transformer networks such as BERT yield state-of-the-art results for text classification tasks.
no code implementations • LREC 2020 • Varvara Logacheva, Denis Teslenko, Artem Shelmanov, Steffen Remus, Dmitry Ustalov, Andrey Kutuzov, Ekaterina Artemova, Chris Biemann, Simone Paolo Ponzetto, Alexander Panchenko
We use this method to induce a collection of sense inventories for 158 languages on the basis of the original pre-trained fastText word embeddings by Grave et al. (2018), enabling WSD in these languages.
1 code implementation • LREC 2020 • Seid Muhie Yimam, Gopalakrishnan Venkatesh, John Sie Yuen Lee, Chris Biemann
The aim is to build a writing aid system that automatically edits a text so that it better adheres to the academic style of writing.
no code implementations • 9 Dec 2019 • Seid Muhie Yimam, Abinew Ali Ayele, Chris Biemann
Since several languages can be written using the Fidel script, we have used the existing Amharic, Tigrinya and Ge'ez corpora to retain only the Amharic tweets.
1 code implementation • 23 Sep 2019 • Gregor Wiedemann, Steffen Remus, Avi Chawla, Chris Biemann
Since vectors of the same word type can vary depending on the respective context, they implicitly provide a model for word sense disambiguation (WSD).
1 code implementation • ACL 2019 • Artem Chernodub, Oleksiy Oliynyk, Philipp Heidenreich, Alex Bondarenko, Matthias Hagen, Chris Biemann, Alex Panchenko, er
We present TARGER, an open source neural argument mining framework for tagging arguments in free input texts and for keyword-based retrieval of arguments from an argument-tagged web-scale corpus.
no code implementations • ACL 2019 • {\"O}zge Sevgili, Alex Panchenko, er, Chris Biemann
Entity Disambiguation (ED) is the task of linking an ambiguous entity mention to a corresponding entry in a knowledge base.
1 code implementation • ACL 2019 • Rami Aly, Steffen Remus, Chris Biemann
Capsule networks have been shown to demonstrate good performance on structured data in the area of visual inference.
no code implementations • ACL 2019 • Abhik Jana, Dima Puzyrev, Alex Panchenko, er, Pawan Goyal, Chris Biemann, Animesh Mukherjee
In particular, we use hypernymy information of the multiword and its constituents encoded in the form of the recently introduced Poincar{\'e} embeddings in addition to the distributional information to detect compositionality for noun phrases.
1 code implementation • ACL 2019 • Andrey Kutuzov, Mohammad Dorgham, Oleksiy Oliynyk, Chris Biemann, Alexander Panchenko
The computation of distance measures between nodes in graphs is inefficient and does not scale to large graphs.
1 code implementation • ACL 2019 • Max Friedrich, Arne Köhn, Gregor Wiedemann, Chris Biemann
De-identification is the task of detecting protected health information (PHI) in medical text.
no code implementations • 7 Jun 2019 • Abhik Jana, Dmitry Puzyrev, Alexander Panchenko, Pawan Goyal, Chris Biemann, Animesh Mukherjee
In particular, we use hypernymy information of the multiword and its constituents encoded in the form of the recently introduced Poincar\'e embeddings in addition to the distributional information to detect compositionality for noun phrases.
1 code implementation • ACL 2019 • Rami Aly, Shantanu Acharya, Alexander Ossa, Arne Köhn, Chris Biemann, Alexander Panchenko
We introduce the use of Poincar\'e embeddings to improve existing state-of-the-art approaches to domain-specific taxonomy induction from text as a signal for both relocating wrong hyponym terms within a (pre-induced) taxonomy as well as for attaching disconnected terms in a taxonomy.
no code implementations • WS 2019 • Dirk Johann{\ss}en, Chris Biemann, David Scheffer
In addition, we found a significant correlation of r = . 2 between subsequent academic success and data automatically labeled with our model in an extrinsic evaluation.
1 code implementation • NAACL 2019 • Tim Fischer, Steffen Remus, Chris Biemann
Particularly for dynamic systems, where topics are not predefined but formulated as a search query, we believe a more informative approach is to perform user studies for directly comparing different methods in the same view.
no code implementations • SEMEVAL 2019 • Gregor Wiedemann, Eugen Ruppert, Chris Biemann
We present a neural network based approach of transfer learning for offensive language detection.
1 code implementation • SEMEVAL 2019 • Saba Anwar, Dmitry Ustalov, Nikolay Arefyev, Simone Paolo Ponzetto, Chris Biemann, Alexander Panchenko
We present our system for semantic frame induction that showed the best performance in Subtask B. 1 and finished as the runner-up in Subtask A of the SemEval 2019 Task 2 on unsupervised semantic frame induction (QasemiZadeh et al., 2019).
no code implementations • WS 2019 • Md. Shad Akhtar, Abhishek Kumar, Asif Ekbal, Chris Biemann, Pushpak Bhattacharyya
In this paper, we propose a language-agnostic deep neural network architecture for aspect-based sentiment analysis.
Aspect-Based Sentiment Analysis Aspect-Based Sentiment Analysis (ABSA) +3
no code implementations • 15 Jan 2019 • Matthias Schildwächter, Alexander Bondarenko, Julian Zenker, Matthias Hagen, Chris Biemann, Alexander Panchenko
We present CAM (comparative argumentative machine), a novel open-domain IR system to argumentatively compare objects with respect to information extracted from the Common Crawl.
no code implementations • 7 Nov 2018 • Gregor Wiedemann, Eugen Ruppert, Raghav Jindal, Chris Biemann
Best results are achieved from pre-training our model on the unsupervised topic clustering of tweets in combination with thematic user cluster information.
no code implementations • 7 Nov 2018 • Gregor Wiedemann, Raghav Jindal, Chris Biemann
We evaluate the performance of different word and character embeddings on two standard German datasets and with a special focus on out-of-vocabulary words.
3 code implementations • WS 2019 • Alexander Panchenko, Alexander Bondarenko, Mirco Franzek, Matthias Hagen, Chris Biemann
We tackle the tasks of automatically identifying comparative sentences and categorizing the intended preference (e. g., "Python has better NLP libraries than MATLAB" => (Python, better, MATLAB).
1 code implementation • 17 Sep 2018 • Dmitry Ustalov, Alexander Panchenko, Chris Biemann, Simone Paolo Ponzetto
In this paper, we show how unsupervised sense representations can be used to improve hypernymy extraction.
no code implementations • EMNLP 2018 • Gregor Wiedemann, Seid Muhie Yimam, Chris Biemann
We introduce an advanced information extraction pipeline to automatically process very large collections of unstructured textual data for the purpose of investigative journalism.
no code implementations • CL 2018 • Martin Riedl, Chris Biemann
First, we introduce DRUID, which is a method for detecting MWEs.
no code implementations • EMNLP 2018 • Seid Muhie Yimam, Chris Biemann
In this paper, we present Par4Sem, a semantic writing aid tool based on adaptive paraphrasing.
2 code implementations • CL 2019 • Dmitry Ustalov, Alexander Panchenko, Chris Biemann, Simone Paolo Ponzetto
We present a detailed theoretical and computational analysis of the Watset meta-algorithm for fuzzy graph clustering, which has been found to be widely applicable in a variety of domains.
no code implementations • SEMEVAL 2019 • Andrey Kutuzov, Mohammad Dorgham, Oleksiy Oliynyk, Chris Biemann, Alexander Panchenko
We present path2vec, a new approach for learning graph embeddings that relies on structural measures of pairwise node similarities.
no code implementations • 13 Jul 2018 • Gregor Wiedemann, Seid Muhie Yimam, Chris Biemann
Investigative journalism in recent years is confronted with two major challenges: 1) vast amounts of unstructured data originating from large text collections such as leaks or answers to Freedom of Information requests, and 2) multi-lingual data due to intensified global cooperation and communication in politics, business and civil society.
no code implementations • COLING 2018 • Seid Muhie Yimam, Chris Biemann
Learning from a real-world data stream and continuously updating the model without explicit supervision is a new challenge for NLP applications with machine learning components.
no code implementations • NAACL 2018 • Ahmed Elsafty, Martin Riedl, Chris Biemann
Detecting the similarity between job advertisements is important for job recommendation systems as it allows, for example, the application of item-to-item based recommendations.
1 code implementation • ACL 2018 • Dmitry Ustalov, Alexander Panchenko, Andrei Kutuzov, Chris Biemann, Simone Paolo Ponzetto
We use dependency triples automatically extracted from a Web-scale corpus to perform unsupervised semantic frame induction.
no code implementations • SEMEVAL 2018 • Enrico Santus, Chris Biemann, Emmanuele Chersoni
This paper describes BomJi, a supervised system for capturing discriminative attributes in word pairs (e. g. yellow as discriminative for banana over watermelon).
Ranked #3 on Relation Extraction on SemEval 2018 Task 10
1 code implementation • LREC 2018 • Dmitry Ustalov, Denis Teslenko, Alexander Panchenko, Mikhail Chernoskutov, Chris Biemann, Simone Paolo Ponzetto
The sparse mode uses the traditional vector space model to estimate the most similar word sense corresponding to its context.
no code implementations • WS 2018 • Seid Muhie Yimam, Chris Biemann, Shervin Malmasi, Gustavo H. Paetzold, Lucia Specia, Sanja Štajner, Anaïs Tack, Marcos Zampieri
We report the findings of the second Complex Word Identification (CWI) shared task organized as part of the BEA workshop co-located with NAACL-HLT'2018.
no code implementations • 18 Apr 2018 • Benjamin Milde, Chris Biemann
We introduce "Unspeech" embeddings, which are based on unsupervised learning of context feature representations for spoken language.
no code implementations • LREC 2018 • Stefano Faralli, Alexander Panchenko, Chris Biemann, Simone Paolo Ponzetto
We introduce a new lexical resource that enriches the Framester knowledge graph, which links Framnet, WordNet, VerbNet and other resources, with semantic features from text corpora.
no code implementations • 28 Dec 2017 • Andreas Holzinger, Chris Biemann, Constantinos S. Pattichis, Douglas B. Kell
In this paper we outline some of our research topics in the context of the relatively new area of explainable-AI with a focus on the application in medicine, which is a very special domain.
no code implementations • 23 Dec 2017 • Chris Biemann, Stefano Faralli, Alexander Panchenko, Simone Paolo Ponzetto
While both kinds of semantic resources are available with high lexical coverage, our aligned resource combines the domain specificity and availability of contextual information from distributional models with the conciseness and high quality of manually crafted lexical networks.
1 code implementation • LREC 2018 • Alexander Panchenko, Dmitry Ustalov, Stefano Faralli, Simone P. Ponzetto, Chris Biemann
In this paper, we show how distributionally-induced semantic classes can be helpful for extracting hypernyms.
no code implementations • IJCNLP 2017 • Seid Muhie Yimam, Sanja {\v{S}}tajner, Martin Riedl, Chris Biemann
Complex word identification (CWI) is an important task in text accessibility.
no code implementations • LREC 2018 • Alexander Panchenko, Eugen Ruppert, Stefano Faralli, Simone Paolo Ponzetto, Chris Biemann
We present DepCC, the largest-to-date linguistically analyzed corpus in English including 365 million documents, composed of 252 billion tokens and 7. 5 billion of named entity occurrences in 14. 3 billion sentences from a web-scale crawl of the \textsc{Common Crawl} project.
no code implementations • RANLP 2017 • Seid Muhie Yimam, Steffen Remus, Alex Panchenko, er, Andreas Holzinger, Chris Biemann
In this paper, we describe the concept of entity-centric information access for the biomedical domain.
no code implementations • RANLP 2017 • Seid Muhie Yimam, Sanja {\v{S}}tajner, Martin Riedl, Chris Biemann
Complex Word Identification (CWI) is an important task in lexical simplification and text accessibility.
no code implementations • 31 Aug 2017 • Alexander Panchenko, Dmitry Ustalov, Nikolay Arefyev, Denis Paperno, Natalia Konstantinova, Natalia Loukachevitch, Chris Biemann
On the one hand, humans easily make judgments about semantic relatedness.
no code implementations • 30 Aug 2017 • Dmitry Ustalov, Mikhail Chernoskutov, Chris Biemann, Alexander Panchenko
Graph-based synset induction methods, such as MaxMax and Watset, induce synsets by performing a global clustering of a synonymy graph.
1 code implementation • WS 2016 • Maria Pelevina, Nikolay Arefyev, Chris Biemann, Alexander Panchenko
We present a simple yet effective approach for learning word sense embeddings.
1 code implementation • SEMEVAL 2017 • N, Titas i, Chris Biemann, Seid Muhie Yimam, Deepak Gupta, Sarah Kohail, Asif Ekbal, Pushpak Bhattacharyya
In this paper we present the system for Answer Selection and Ranking in Community Question Answering, which we build as part of our participation in SemEval-2017 Task 3.
no code implementations • SEMEVAL 2017 • Abhishek Kumar, Abhishek Sethi, Md. Shad Akhtar, Asif Ekbal, Chris Biemann, Pushpak Bhattacharyya
The other system was based on Support Vector Regression using word embeddings, lexicon features, and PMI scores as features.
no code implementations • SEMEVAL 2017 • Sarah Kohail, Amr Rekaby Salama, Chris Biemann
This paper reports the STS-UHH participation in the SemEval 2017 shared Task 1 of Semantic Textual Similarity (STS).
1 code implementation • EMNLP 2017 • Alexander Panchenko, Fide Marten, Eugen Ruppert, Stefano Faralli, Dmitry Ustalov, Simone Paolo Ponzetto, Chris Biemann
In word sense disambiguation (WSD), knowledge-based systems tend to be much more interpretable than knowledge-free counterparts as they rely on the wealth of manually-encoded elements representing word senses, such as hypernyms, usage examples, and images.
1 code implementation • EACL 2017 • Dmitry Ustalov, Nikolay Arefyev, Chris Biemann, Alexander Panchenko
We present a new approach to extraction of hypernyms based on projection learning and word embeddings.
1 code implementation • ACL 2017 • Dmitry Ustalov, Alexander Panchenko, Chris Biemann
This paper presents a new graph-based approach that induces synsets using synonymy dictionaries and word embeddings.
no code implementations • EACL 2017 • Stefano Faralli, Alex Panchenko, er, Chris Biemann, Simone Paolo Ponzetto
In this paper, we present ContrastMedium, an algorithm that transforms noisy semantic networks into full-fledged, clean taxonomies.
no code implementations • EACL 2017 • Alex Panchenko, er, Eugen Ruppert, Stefano Faralli, Simone Paolo Ponzetto, Chris Biemann
On the example of word sense induction and disambiguation (WSID), we show that it is possible to develop an interpretable model that matches the state-of-the-art models in accuracy.
no code implementations • WS 2017 • Alex Panchenko, er, Stefano Faralli, Simone Paolo Ponzetto, Chris Biemann
We introduce a new method for unsupervised knowledge-based word sense disambiguation (WSD) based on a resource that links two types of sense-aware lexical networks: one is induced from a corpus using distributional semantics, the other is manually constructed.
no code implementations • WS 2016 • Michael Zock, Chris Biemann
To this end, we asked crowdworkers to provide some cues to describe a given target and to specify then how each one of them relates to the target, in the hope that this could help others to find the elusive word.
no code implementations • COLING 2016 • Benjamin Milde, Jonas Wacker, Stefan Radomski, Max M{\"u}hlh{\"a}user, Chris Biemann
In this demonstration paper we describe Ambient Search, a system that displays and retrieves documents in real time based on speech input.
no code implementations • COLING 2016 • Benjamin Milde, Jonas Wacker, Stefan Radomski, Max M{\"u}hlh{\"a}user, Chris Biemann
We present Ambient Search, an open source system for displaying and retrieving relevant documents in real time for speech input.
no code implementations • WS 2016 • Chris Biemann
Distributional Semantic Models (DSMs) have recently received increased attention, together with the rise of neural architectures for scalable training of dense vector embeddings.
no code implementations • WS 2016 • Richard Eckart de Castilho, {\'E}va M{\'u}jdricza-Maydt, Seid Muhie Yimam, Silvana Hartmann, Iryna Gurevych, Anette Frank, Chris Biemann
We introduce the third major release of WebAnno, a generic web-based annotation tool for distributed teams.
no code implementations • ACL 2016 • Seid Muhie Yimam, Heiner Ulrich, von L, Tatiana esberger, Marcel Rosenbach, Michaela Regneri, Alex Panchenko, er, Franziska Lehmann, Uli Fahrer, Chris Biemann, Kathrin Ballweg
no code implementations • SEMEVAL 2016 • Alex Panchenko, er, Stefano Faralli, Eugen Ruppert, Steffen Remus, Hubert Naets, C{\'e}drick Fairon, Simone Paolo Ponzetto, Chris Biemann
no code implementations • SEMEVAL 2016 • Ayush Kumar, Sarah Kohail, Amit Kumar, Asif Ekbal, Chris Biemann
Aspect-Based Sentiment Analysis Aspect-Based Sentiment Analysis (ABSA) +1
no code implementations • LREC 2016 • Steffen Remus, Chris Biemann
This work presents a straightforward method for extending or creating in-domain web corpora by focused webcrawling.
no code implementations • LREC 2016 • Darina Benikova, Chris Biemann
Semantic relations play an important role in linguistic knowledge representation.
no code implementations • ACL 2014 • Sunny Mitra, Ritwik Mitra, Martin Riedl, Chris Biemann, Animesh Mukherjee, Pawan Goyal
In this paper, we propose an unsupervised method to identify noun sense changes based on rigorous analysis of time-varying text data available in the form of millions of digitized books.
no code implementations • LREC 2014 • Kostadin Cholakov, Chris Biemann, Judith Eckle-Kohler, Iryna Gurevych
This article describes a lexical substitution dataset for German.
no code implementations • LREC 2014 • Darina Benikova, Chris Biemann, Marc Reznicek
We describe our approach to creating annotation guidelines based on linguistic and semantic considerations, and how we iteratively refined and tested them in the early stages of annotation in order to arrive at the largest publicly available dataset for German NER, consisting of over 31, 000 manually annotated sentences (over 591, 000 tokens) from German Wikipedia and German online news.
no code implementations • LREC 2014 • Martin Riedl, Richard Steuer, Chris Biemann
This paper introduces a distributional thesaurus and sense clusters computed on the complete Google Syntactic N-grams, which is extracted from Google Books, a very large corpus of digitized books published between 1520 and 2008.
no code implementations • LREC 2012 • Chris Biemann
This lexical resource, created by a crowdsourcing process using Amazon Mechanical Turk (http://www. mturk. com), encompasses a sense inventory for lexical substitution for 1, 012 highly frequent English common nouns.