no code implementations • EMNLP 2020 • Philipp Dufter, Hinrich Sch{\"u}tze
We aim to identify architectural properties of BERT and linguistic properties of languages that are necessary for BERT to become multilingual.
no code implementations • Findings (EMNLP) 2021 • Antonis Maronikolakis, Philipp Dufter, Hinrich Schütze
The size of the vocabulary is a central design choice in large pretrained language models, with respect to both performance and memory requirements.
1 code implementation • Findings (NAACL) 2022 • Victor Steinborn, Philipp Dufter, Haris Jabbar, Hinrich Schuetze
Bias research in NLP is a rapidly growing and developing field.
no code implementations • 11 Apr 2024 • Haotian Zhang, Haoxuan You, Philipp Dufter, BoWen Zhang, Chen Chen, Hong-You Chen, Tsu-Jui Fu, William Yang Wang, Shih-Fu Chang, Zhe Gan, Yinfei Yang
While Ferret seamlessly integrates regional understanding into the Large Language Model (LLM) to facilitate its referring and grounding capability, it poses certain limitations: constrained by the pre-trained fixed visual encoder and failed to perform well on broader tasks.
Ranked #83 on Visual Question Answering on MM-Vet
no code implementations • 14 Mar 2024 • Brandon McKinzie, Zhe Gan, Jean-Philippe Fauconnier, Sam Dodge, BoWen Zhang, Philipp Dufter, Dhruti Shah, Xianzhi Du, Futang Peng, Floris Weers, Anton Belyi, Haotian Zhang, Karanjeet Singh, Doug Kang, Ankur Jain, Hongyu Hè, Max Schwarzer, Tom Gunter, Xiang Kong, Aonan Zhang, Jianyu Wang, Chong Wang, Nan Du, Tao Lei, Sam Wiseman, Guoli Yin, Mark Lee, ZiRui Wang, Ruoming Pang, Peter Grasch, Alexander Toshev, Yinfei Yang
Through careful and comprehensive ablations of the image encoder, the vision language connector, and various pre-training data choices, we identified several crucial design lessons.
Ranked #30 on Visual Question Answering on MM-Vet
no code implementations • LREC 2022 • Silvia Severini, Ayyoob Imani, Philipp Dufter, Hinrich Schütze
Prior work on extracting MNE datasets from parallel corpora required resources such as large monolingual corpora or word aligners that are unavailable or perform poorly for underresourced languages.
no code implementations • EMNLP (insights) 2021 • Antonis Maronikolakis, Philipp Dufter, Hinrich Schütze
We show that the closer two languages are, the better BERT can align them on the character level.
1 code implementation • 16 Sep 2021 • Sheng Liang, Philipp Dufter, Hinrich Schütze
Multilingual pretrained language models (MPLMs) exhibit multilinguality and are well suited for transfer across languages.
no code implementations • 13 Sep 2021 • Antonis Maronikolakis, Philipp Dufter, Hinrich Schütze
The size of the vocabulary is a central design choice in large pretrained language models, with respect to both performance and memory requirements.
1 code implementation • EMNLP 2021 • Ayyoob Imani, Masoud Jalili Sabet, Lütfi Kerem Şenel, Philipp Dufter, François Yvon, Hinrich Schütze
With the advent of end-to-end deep learning approaches in machine translation, interest in word alignments initially decreased; however, they have again become a focus of research more recently.
no code implementations • ACL 2021 • Ayyoob Imani, Masoud Jalili Sabet, Philipp Dufter, Michael Cysouw, Hinrich Schütze
With more than 7000 languages worldwide, multilingual natural language processing (NLP) is essential both from an academic and commercial perspective.
1 code implementation • NAACL 2021 • Philipp Dufter, Nora Kassner, Hinrich Schütze
Recent research investigates factual knowledge stored in large pretrained language models (PLMs).
no code implementations • CL (ACL) 2022 • Philipp Dufter, Martin Schmitt, Hinrich Schütze
Transformers are arguably the main workhorse in recent Natural Language Processing research.
1 code implementation • EACL 2021 • Nora Kassner, Philipp Dufter, Hinrich Schütze
(i) Can mBERT be used as a multilingual knowledge base?
no code implementations • 21 Dec 2020 • Ehsaneddin Asgari, Masoud Jalili Sabet, Philipp Dufter, Christopher Ringlstetter, Hinrich Schütze
This method's hypothesis is that the aggregation of different granularities of text for certain language pairs can help word-level alignment.
1 code implementation • COLING 2020 • Philipp Dufter, Martin Schmitt, Hinrich Sch{\"u}tze
Self-Attention Networks (SANs) are an integral part of successful neural architectures such as Transformer (Vaswani et al., 2017), and thus of pretrained language models such as BERT (Devlin et al., 2019) or GPT-3 (Brown et al., 2020).
1 code implementation • COLING 2020 • Sheng Liang, Philipp Dufter, Hinrich Sch{\"u}tze
Pretrained language models (PLMs) learn stereotypes held by humans and reflected in text from their training corpora, including gender bias.
no code implementations • NAACL (TextGraphs) 2021 • Martin Schmitt, Leonardo F. R. Ribeiro, Philipp Dufter, Iryna Gurevych, Hinrich Schütze
We present Graformer, a novel Transformer-based encoder-decoder architecture for graph-to-text generation.
Ranked #5 on KG-to-Text Generation on AGENDA
1 code implementation • 1 May 2020 • Philipp Dufter, Hinrich Schütze
We aim to identify architectural properties of BERT and linguistic properties of languages that are necessary for BERT to become multilingual.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Mengjie Zhao, Philipp Dufter, Yadollah Yaghoobzadeh, Hinrich Schütze
Pretrained language models have achieved a new state of the art on many NLP tasks, but there are still many open questions about how and why they work so well.
3 code implementations • Findings of the Association for Computational Linguistics 2020 • Masoud Jalili Sabet, Philipp Dufter, François Yvon, Hinrich Schütze
We find that alignments created from embeddings are superior for four and comparable for two language pairs compared to those produced by traditional statistical aligners, even with abundant parallel data; e. g., contextualized embeddings achieve a word alignment F1 for English-German that is 5 percentage points higher than eflomal, a high-quality statistical aligner, trained on 100k parallel sentences.
1 code implementation • IJCNLP 2019 • Philipp Dufter, Hinrich Schütze
In this work, we investigate three methods for making word spaces interpretable by rotation: Densifier (Rothe et al., 2016), linear SVMs and DensRay, a new method we propose.
no code implementations • 1 Nov 2018 • Philipp Dufter, Mengjie Zhao, Hinrich Schütze
A simple and effective context-based multilingual embedding learner is Levy et al. (2017)'s S-ID (sentence ID) method.
no code implementations • ACL 2018 • Philipp Dufter, Mengjie Zhao, Martin Schmitt, Alexander Fraser, Hinrich Schütze
We present a new method for estimating vector space representations of words: embedding learning by concept induction.