1 code implementation • 20 Sep 2023 • Atakan Kara, Farrin Marouf Sofian, Andrew Bond, Gözde Gül Şahin
To encourage further research on Turkish GEC, we release our datasets, baseline models, and the synthetic data generation pipeline at https://github. com/GGLAB-KU/gecturk.
1 code implementation • 13 Sep 2023 • Arda Uzunoğlu, Gözde Gül Şahin
To tackle these tasks, we implement strong baseline models via fine-tuning large language-specific models such as TR-BART and BERTurk, as well as multilingual models such as mBART, mT5, and XLM.
1 code implementation • 27 Jul 2023 • Subha Vadlamannati, Gözde Gül Şahin
However, determining the best method to select examples for ICL is nontrivial as the results can vary greatly depending on the quality, quantity, and order of examples used.
Ranked #1 on
Text Simplification
on ASSET
1 code implementation • 25 Apr 2023 • Jan-Christoph Klie, Ji-Ung Lee, Kevin Stowe, Gözde Gül Şahin, Nafise Sadat Moosavi, Luke Bates, Dominic Petrak, Richard Eckart de Castilho, Iryna Gurevych
Citizen Science is an alternative to crowdsourcing that is relatively unexplored in the context of NLP.
1 code implementation • 3 Nov 2022 • Emre Can Acikgoz, Tilek Chubakov, Müge Kural, Gözde Gül Şahin, Deniz Yuret
While transformer architectures with data augmentation achieved the most promising results for inflection and reinflection tasks, prefix-tuning on mGPT received the highest results for the analysis task.
1 code implementation • ACL 2022 • Tim Baumgärtner, Kexin Wang, Rachneet Sachdeva, Max Eichler, Gregor Geigle, Clifton Poth, Hannah Sterz, Haritz Puerto, Leonardo F. R. Ribeiro, Jonas Pfeiffer, Nils Reimers, Gözde Gül Şahin, Iryna Gurevych
Recent advances in NLP and information retrieval have given rise to a diverse set of question answering tasks that are of different formats (e. g., extractive, abstractive), require different model architectures (e. g., generative, discriminative), and setups (e. g., with or without retrieval).
1 code implementation • 3 Dec 2021 • Haritz Puerto, Gözde Gül Şahin, Iryna Gurevych
The recent explosion of question answering (QA) datasets and models has increased the interest in the generalization of models across multiple domains and formats by either training on multiple datasets or by combining multiple models.
no code implementations • CL (ACL) 2022 • Gözde Gül Şahin
Although NLP has recently witnessed a load of textual augmentation techniques, the field still lacks a systematic performance analysis on a diverse set of languages and sequence tagging tasks.
no code implementations • ACL 2020 • Gözde Gül Şahin, Yova Kementchedjhieva, Phillip Rust, Iryna Gurevych
To expose this problem in a new light, we introduce a challenge on learning from small data, PuzzLing Machines, which consists of Rosetta Stone puzzles from Linguistic Olympiads for high school students.
no code implementations • 11 Dec 2019 • Gözde Gül Şahin, Iryna Gurevych
We show that they are able to recover the morphological input parameters, i. e., predicting the lemma (e. g., cat) or the morphological tags (e. g., Plural) when run in the reverse direction, without any significant performance drop in the forward direction, i. e., predicting the surface form (e. g., cats).
1 code implementation • IJCNLP 2019 • Max Eichler, Gözde Gül Şahin, Iryna Gurevych
We present LINSPECTOR WEB, an open source multilingual inspector to analyze word representations.
1 code implementation • NAACL 2019 • Steffen Eger, Gözde Gül Şahin, Andreas Rücklé, Ji-Ung Lee, Claudia Schulz, Mohsen Mesgar, Krishnkant Swarnkar, Edwin Simpson, Iryna Gurevych
Visual modifications to text are often used to obfuscate offensive comments in social media (e. g., "! d10t") or as a writing style ("1337" in "leet speak"), among other scenarios.
3 code implementations • CL 2020 • Gözde Gül Şahin, Clara Vania, Ilia Kuznetsov, Iryna Gurevych
We present a reusable methodology for creation and evaluation of such tests in a multilingual setting.
2 code implementations • EMNLP 2018 • Gözde Gül Şahin, Mark Steedman
Neural NLP systems achieve high scores in the presence of sizable training dataset.
1 code implementation • ACL 2018 • Gözde Gül Şahin, Mark Steedman
Character-level models have become a popular approach specially for their accessibility and ability to handle unseen data.