no code implementations • EMNLP (CODI) 2020 • Haixia Chai, Wei Zhao, Steffen Eger, Michael Strube
A substantial overlap of coreferent mentions in the CoNLL dataset magnifies the recent progress on coreference resolution.
no code implementations • COLING (CODI, CRAC) 2022 • Juntao Yu, Sopan Khosla, Ramesh Manuvinakurike, Lori Levin, Vincent Ng, Massimo Poesio, Michael Strube, Carolyn Rosé
The CODI-CRAC 2022 Shared Task on Anaphora Resolution in Dialogues is the second edition of an initiative focused on detecting different types of anaphoric relations in conversations of different kinds.
1 code implementation • NAACL 2022 • Haixia Chai, Michael Strube
In recent years, transformer-based coreference resolution systems have achieved remarkable improvements on the CoNLL dataset.
1 code implementation • ACL 2022 • Sungho Jeon, Michael Strube
We evaluate our model on three downstream tasks showing that it is not only linguistically more sound than previous models but also that it outperforms them in end applications.
1 code implementation • COLING (CRAC) 2022 • Haixia Chai, Nafise Sadat Moosavi, Iryna Gurevych, Michael Strube
The results of our extrinsic evaluation show that while there is a significant difference between the performance of the rule-based system vs. state-of-the-art neural model on coreference resolution datasets, we do not observe a considerable difference on their impact on downstream models.
1 code implementation • EMNLP (newsum) 2021 • Mehwish Fatima, Michael Strube
We collect high-quality, real-world cross-lingual data from Spektrum der Wissenschaft, which publishes human-written German scientific summaries of English science articles on various subjects.
no code implementations • EMNLP 2020 • Sungho Jeon, Michael Strube
We approximate a linguistic theory of coherence, Centering theory, which we use to track the changes of focus between discourse segments.
no code implementations • EMNLP (sdp) 2020 • Mark-Christoph Müller, Sucheta Ghosh, Maja Rey, Ulrike Wittig, Wolfgang Müller, Michael Strube
We introduce a novel scientific document processing task for making previously inaccessible information in printed paper documents available to automatic processing.
1 code implementation • EMNLP (sustainlp) 2021 • Sungho Jeon, Michael Strube
In this work, we first show that state-of-the-art systems, recent neural essay scoring systems, might be also influenced by the correlation between essay length and scores in a standard dataset.
Ranked #3 on Automated Essay Scoring on ASAP-AES
no code implementations • NAACL (ClinicalNLP) 2022 • Siting Liang, Klaus Kades, Matthias Fink, Peter Full, Tim Weber, Jens Kleesiek, Michael Strube, Klaus Maier-Hein
Writing the conclusion section of radiology reports is essential for communicating the radiology findings and its assessment to physician in a condensed form.
no code implementations • ACL (CODI, CRAC) 2021 • Sopan Khosla, Juntao Yu, Ramesh Manuvinakurike, Vincent Ng, Massimo Poesio, Michael Strube, Carolyn Rosé
In this paper, we provide an overview of the CODI-CRAC 2021 Shared-Task: Anaphora Resolution in Dialogue.
1 code implementation • 1 Apr 2024 • Wei Liu, Stephen Wan, Michael Strube
We consider an unanswered question in the discourse processing community: why do relation classifiers trained on explicit examples (with connectives removed) perform poorly in real implicit scenarios?
1 code implementation • 1 Feb 2024 • Xianghe Ma, Michael Strube, Wei Zhao
To address this issue, we propose a graph-based clustering approach to capture nuanced changes in both high- and low-frequency word senses across time and languages, including the acquisition and loss of these senses over time.
1 code implementation • 3 Dec 2023 • Diaaeldin Taha, Wei Zhao, J. Maxwell Riestenberg, Michael Strube
Theoretical results from discrete geometry suggest that normed spaces can abstractly embed finite metric spaces with surprisingly low theoretical bounds on distortion in low dimensions.
1 code implementation • 26 Oct 2023 • Haixia Chai, Michael Strube
Multilingual coreference resolution (MCR) has been a long-standing and challenging task.
1 code implementation • 24 Jun 2023 • Wei Zhao, Federico Lopez, J. Maxwell Riestenberg, Michael Strube, Diaaeldin Taha, Steve Trettel
The uniform geometry of Euclidean and hyperbolic spaces allows for representing graphs with uniform geometric and topological features, such as grids and hierarchies, with minimal distortion.
1 code implementation • 10 Jun 2023 • Wei Liu, Xiyan Fu, Michael Strube
Coherence is an important aspect of text quality, and various approaches have been applied to coherence modeling.
1 code implementation • 10 Jun 2023 • Wei Liu, Michael Strube
Implicit discourse relation classification is a challenging task due to the absence of discourse connectives.
no code implementations • 4 Apr 2023 • Mehwish Fatima, Tim Kolber, Katja Markert, Michael Strube
Cross-lingual science journalism generates popular science stories of scientific articles different from the source language for a non-expert audience.
Abstractive Text Summarization Cross-Lingual Abstractive Summarization
4 code implementations • 9 Jun 2022 • Aarohi Srivastava, Abhinav Rastogi, Abhishek Rao, Abu Awal Md Shoeb, Abubakar Abid, Adam Fisch, Adam R. Brown, Adam Santoro, Aditya Gupta, Adrià Garriga-Alonso, Agnieszka Kluska, Aitor Lewkowycz, Akshat Agarwal, Alethea Power, Alex Ray, Alex Warstadt, Alexander W. Kocurek, Ali Safaya, Ali Tazarv, Alice Xiang, Alicia Parrish, Allen Nie, Aman Hussain, Amanda Askell, Amanda Dsouza, Ambrose Slone, Ameet Rahane, Anantharaman S. Iyer, Anders Andreassen, Andrea Madotto, Andrea Santilli, Andreas Stuhlmüller, Andrew Dai, Andrew La, Andrew Lampinen, Andy Zou, Angela Jiang, Angelica Chen, Anh Vuong, Animesh Gupta, Anna Gottardi, Antonio Norelli, Anu Venkatesh, Arash Gholamidavoodi, Arfa Tabassum, Arul Menezes, Arun Kirubarajan, Asher Mullokandov, Ashish Sabharwal, Austin Herrick, Avia Efrat, Aykut Erdem, Ayla Karakaş, B. Ryan Roberts, Bao Sheng Loe, Barret Zoph, Bartłomiej Bojanowski, Batuhan Özyurt, Behnam Hedayatnia, Behnam Neyshabur, Benjamin Inden, Benno Stein, Berk Ekmekci, Bill Yuchen Lin, Blake Howald, Bryan Orinion, Cameron Diao, Cameron Dour, Catherine Stinson, Cedrick Argueta, César Ferri Ramírez, Chandan Singh, Charles Rathkopf, Chenlin Meng, Chitta Baral, Chiyu Wu, Chris Callison-Burch, Chris Waites, Christian Voigt, Christopher D. Manning, Christopher Potts, Cindy Ramirez, Clara E. Rivera, Clemencia Siro, Colin Raffel, Courtney Ashcraft, Cristina Garbacea, Damien Sileo, Dan Garrette, Dan Hendrycks, Dan Kilman, Dan Roth, Daniel Freeman, Daniel Khashabi, Daniel Levy, Daniel Moseguí González, Danielle Perszyk, Danny Hernandez, Danqi Chen, Daphne Ippolito, Dar Gilboa, David Dohan, David Drakard, David Jurgens, Debajyoti Datta, Deep Ganguli, Denis Emelin, Denis Kleyko, Deniz Yuret, Derek Chen, Derek Tam, Dieuwke Hupkes, Diganta Misra, Dilyar Buzan, Dimitri Coelho Mollo, Diyi Yang, Dong-Ho Lee, Dylan Schrader, Ekaterina Shutova, Ekin Dogus Cubuk, Elad Segal, Eleanor Hagerman, Elizabeth Barnes, Elizabeth Donoway, Ellie Pavlick, Emanuele Rodola, Emma Lam, Eric Chu, Eric Tang, Erkut Erdem, Ernie Chang, Ethan A. Chi, Ethan Dyer, Ethan Jerzak, Ethan Kim, Eunice Engefu Manyasi, Evgenii Zheltonozhskii, Fanyue Xia, Fatemeh Siar, Fernando Martínez-Plumed, Francesca Happé, Francois Chollet, Frieda Rong, Gaurav Mishra, Genta Indra Winata, Gerard de Melo, Germán Kruszewski, Giambattista Parascandolo, Giorgio Mariani, Gloria Wang, Gonzalo Jaimovitch-López, Gregor Betz, Guy Gur-Ari, Hana Galijasevic, Hannah Kim, Hannah Rashkin, Hannaneh Hajishirzi, Harsh Mehta, Hayden Bogar, Henry Shevlin, Hinrich Schütze, Hiromu Yakura, Hongming Zhang, Hugh Mee Wong, Ian Ng, Isaac Noble, Jaap Jumelet, Jack Geissinger, Jackson Kernion, Jacob Hilton, Jaehoon Lee, Jaime Fernández Fisac, James B. Simon, James Koppel, James Zheng, James Zou, Jan Kocoń, Jana Thompson, Janelle Wingfield, Jared Kaplan, Jarema Radom, Jascha Sohl-Dickstein, Jason Phang, Jason Wei, Jason Yosinski, Jekaterina Novikova, Jelle Bosscher, Jennifer Marsh, Jeremy Kim, Jeroen Taal, Jesse Engel, Jesujoba Alabi, Jiacheng Xu, Jiaming Song, Jillian Tang, Joan Waweru, John Burden, John Miller, John U. Balis, Jonathan Batchelder, Jonathan Berant, Jörg Frohberg, Jos Rozen, Jose Hernandez-Orallo, Joseph Boudeman, Joseph Guerr, Joseph Jones, Joshua B. Tenenbaum, Joshua S. Rule, Joyce Chua, Kamil Kanclerz, Karen Livescu, Karl Krauth, Karthik Gopalakrishnan, Katerina Ignatyeva, Katja Markert, Kaustubh D. Dhole, Kevin Gimpel, Kevin Omondi, Kory Mathewson, Kristen Chiafullo, Ksenia Shkaruta, Kumar Shridhar, Kyle McDonell, Kyle Richardson, Laria Reynolds, Leo Gao, Li Zhang, Liam Dugan, Lianhui Qin, Lidia Contreras-Ochando, Louis-Philippe Morency, Luca Moschella, Lucas Lam, Lucy Noble, Ludwig Schmidt, Luheng He, Luis Oliveros Colón, Luke Metz, Lütfi Kerem Şenel, Maarten Bosma, Maarten Sap, Maartje ter Hoeve, Maheen Farooqi, Manaal Faruqui, Mantas Mazeika, Marco Baturan, Marco Marelli, Marco Maru, Maria Jose Ramírez Quintana, Marie Tolkiehn, Mario Giulianelli, Martha Lewis, Martin Potthast, Matthew L. Leavitt, Matthias Hagen, Mátyás Schubert, Medina Orduna Baitemirova, Melody Arnaud, Melvin McElrath, Michael A. Yee, Michael Cohen, Michael Gu, Michael Ivanitskiy, Michael Starritt, Michael Strube, Michał Swędrowski, Michele Bevilacqua, Michihiro Yasunaga, Mihir Kale, Mike Cain, Mimee Xu, Mirac Suzgun, Mitch Walker, Mo Tiwari, Mohit Bansal, Moin Aminnaseri, Mor Geva, Mozhdeh Gheini, Mukund Varma T, Nanyun Peng, Nathan A. Chi, Nayeon Lee, Neta Gur-Ari Krakover, Nicholas Cameron, Nicholas Roberts, Nick Doiron, Nicole Martinez, Nikita Nangia, Niklas Deckers, Niklas Muennighoff, Nitish Shirish Keskar, Niveditha S. Iyer, Noah Constant, Noah Fiedel, Nuan Wen, Oliver Zhang, Omar Agha, Omar Elbaghdadi, Omer Levy, Owain Evans, Pablo Antonio Moreno Casares, Parth Doshi, Pascale Fung, Paul Pu Liang, Paul Vicol, Pegah Alipoormolabashi, Peiyuan Liao, Percy Liang, Peter Chang, Peter Eckersley, Phu Mon Htut, Pinyu Hwang, Piotr Miłkowski, Piyush Patil, Pouya Pezeshkpour, Priti Oli, Qiaozhu Mei, Qing Lyu, Qinlang Chen, Rabin Banjade, Rachel Etta Rudolph, Raefer Gabriel, Rahel Habacker, Ramon Risco, Raphaël Millière, Rhythm Garg, Richard Barnes, Rif A. Saurous, Riku Arakawa, Robbe Raymaekers, Robert Frank, Rohan Sikand, Roman Novak, Roman Sitelew, Ronan LeBras, Rosanne Liu, Rowan Jacobs, Rui Zhang, Ruslan Salakhutdinov, Ryan Chi, Ryan Lee, Ryan Stovall, Ryan Teehan, Rylan Yang, Sahib Singh, Saif M. Mohammad, Sajant Anand, Sam Dillavou, Sam Shleifer, Sam Wiseman, Samuel Gruetter, Samuel R. Bowman, Samuel S. Schoenholz, Sanghyun Han, Sanjeev Kwatra, Sarah A. Rous, Sarik Ghazarian, Sayan Ghosh, Sean Casey, Sebastian Bischoff, Sebastian Gehrmann, Sebastian Schuster, Sepideh Sadeghi, Shadi Hamdan, Sharon Zhou, Shashank Srivastava, Sherry Shi, Shikhar Singh, Shima Asaadi, Shixiang Shane Gu, Shubh Pachchigar, Shubham Toshniwal, Shyam Upadhyay, Shyamolima, Debnath, Siamak Shakeri, Simon Thormeyer, Simone Melzi, Siva Reddy, Sneha Priscilla Makini, Soo-Hwan Lee, Spencer Torene, Sriharsha Hatwar, Stanislas Dehaene, Stefan Divic, Stefano Ermon, Stella Biderman, Stephanie Lin, Stephen Prasad, Steven T. Piantadosi, Stuart M. Shieber, Summer Misherghi, Svetlana Kiritchenko, Swaroop Mishra, Tal Linzen, Tal Schuster, Tao Li, Tao Yu, Tariq Ali, Tatsu Hashimoto, Te-Lin Wu, Théo Desbordes, Theodore Rothschild, Thomas Phan, Tianle Wang, Tiberius Nkinyili, Timo Schick, Timofei Kornev, Titus Tunduny, Tobias Gerstenberg, Trenton Chang, Trishala Neeraj, Tushar Khot, Tyler Shultz, Uri Shaham, Vedant Misra, Vera Demberg, Victoria Nyamai, Vikas Raunak, Vinay Ramasesh, Vinay Uday Prabhu, Vishakh Padmakumar, Vivek Srikumar, William Fedus, William Saunders, William Zhang, Wout Vossen, Xiang Ren, Xiaoyu Tong, Xinran Zhao, Xinyi Wu, Xudong Shen, Yadollah Yaghoobzadeh, Yair Lakretz, Yangqiu Song, Yasaman Bahri, Yejin Choi, Yichi Yang, Yiding Hao, Yifu Chen, Yonatan Belinkov, Yu Hou, Yufang Hou, Yuntao Bai, Zachary Seid, Zhuoye Zhao, Zijian Wang, Zijie J. Wang, ZiRui Wang, Ziyi Wu
BIG-bench focuses on tasks that are believed to be beyond the capabilities of current language models.
1 code implementation • 26 Jan 2022 • Wei Zhao, Michael Strube, Steffen Eger
Still, recent BERT-based evaluation metrics are weak in recognizing coherence, and thus are not reliable in a way to spot the discourse-level improvements of those text generation systems.
no code implementations • 6 Dec 2021 • Kevin Alex Mathews, Michael Strube
We show that the target word is less useful for detecting metonymy in our dataset.
1 code implementation • NeurIPS 2021 • Federico López, Beatrice Pozzetti, Steve Trettel, Michael Strube, Anna Wienhard
We propose the use of the vector-valued distance to compute distances and extract geometric information from the manifold of symmetric positive definite matrices (SPD), and develop gyrovector calculus, constructing analogs of vector space operations in this curved space.
1 code implementation • 20 Sep 2021 • Federico López, Martin Scholz, Jessica Yung, Marie Pellat, Michael Strube, Lucas Dixon
This paper introduces a simple and effective form of data augmentation for recommender systems.
2 code implementations • 9 Jun 2021 • Federico López, Beatrice Pozzetti, Steve Trettel, Michael Strube, Anna Wienhard
We propose the systematic use of symmetric spaces in representation learning, a class encompassing many of the previously used embedding targets.
1 code implementation • COLING 2020 • Sungho Jeon, Michael Strube
We propose a coherence model which interprets sentences incrementally to capture lexical relations between them.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Federico López, Michael Strube
Label inventories for fine-grained entity typing have grown in size and complexity.
no code implementations • LREC 2020 • Kevin Alex Mathews, Michael Strube
We propose a new, labelled, high-quality corpus of location metonymy called WiMCor, which is large in size and has high coverage.
no code implementations • WS 2019 • Ivan Sekulić, Michael Strube
Mental health poses a significant challenge for an individual's well-being.
no code implementations • CONLL 2019 • Yi Zhu, Benjamin Heinzerling, Ivan Vulić, Michael Strube, Roi Reichart, Anna Korhonen
Recent work has validated the importance of subword information for word representation learning.
1 code implementation • ACL 2019 • Nafise Sadat Moosavi, Leo Born, Massimo Poesio, Michael Strube
To address this problem, minimum spans are manually annotated in smaller corpora.
1 code implementation • WS 2019 • Federico López, Benjamin Heinzerling, Michael Strube
How can we represent hierarchical information present in large type inventories for entity typing?
1 code implementation • ACL 2019 • Benjamin Heinzerling, Michael Strube
Pretrained contextual and non-contextual subword embeddings have become available in over 250 languages, allowing massively multilingual NLP.
Ranked #3 on Part-Of-Speech Tagging on UD
no code implementations • EMNLP 2018 • Mohsen Mesgar, Michael Strube
We propose a local coherence model that captures the flow of what semantically connects adjacent sentences in a text.
1 code implementation • COLING 2018 • Mark-Christoph Müller, Michael Strube
We present WOMBAT, a Python tool which supports NLP practitioners in accessing word embeddings from code.
no code implementations • CL 2018 • Yufang Hou, Katja Markert, Michael Strube
The second stage, bridging antecedent selection, finds the antecedents for all predicted bridging anaphors.
no code implementations • IJCNLP 2017 • Alex Judea, Michael Strube
In this paper we investigate the performance of event argument identification.
1 code implementation • LREC 2018 • Benjamin Heinzerling, Michael Strube
We present BPEmb, a collection of pre-trained subword unit embeddings in 275 languages, based on Byte-Pair Encoding (BPE).
no code implementations • WS 2017 • Leo Born, Mohsen Mesgar, Michael Strube
Although coherence is an important aspect of any text generation system, it has received little attention in the context of machine translation (MT) so far.
1 code implementation • EMNLP 2018 • Nafise Sadat Moosavi, Michael Strube
We show that generalization improves only slightly by merely using a set of additional linguistic features.
no code implementations • EMNLP 2017 • Benjamin Heinzerling, Nafise Sadat Moosavi, Michael Strube
Selectional preferences have long been claimed to be essential for coreference resolution.
no code implementations • ACL 2017 • Nafise Sadat Moosavi, Michael Strube
Lexical features are a major source of information in state-of-the-art coreference resolvers.
no code implementations • EACL 2017 • Benjamin Heinzerling, Michael Strube, Chin-Yew Lin
We introduce automatic verification as a post-processing step for entity linking (EL).
no code implementations • WS 2017 • Nafise Sadat Moosavi, Michael Strube
Only a year ago, all state-of-the-art coreference resolvers were using an extensive amount of surface features.
no code implementations • WS 2016 • Anja Summa, Bernd Resch, Michael Strube
Most work in NLP analysing microblogs focuses on textual content thus neglecting temporal and spatial information.
no code implementations • COLING 2016 • Alex Judea, Michael Strube
Event extraction is a difficult information extraction task.
no code implementations • TACL 2015 • Sebastian Martschat, Michael Strube
Machine learning approaches to coreference resolution vary greatly in the modeling of the problem: while early approaches operated on the mention pair level, current research focuses on ranking architectures and antecedent trees.
no code implementations • LREC 2012 • Alex Judea, Vivi Nastase, Michael Strube
This paper describes the derivation of distributional semantic representations for open class words relative to a concept inventory, and of concepts relative to open class words through grammatical relations extracted from Wikipedia articles.