no code implementations • 3 Feb 2023 • Ismail Nejjar, Fabian Geissmann, Mengjie Zhao, Cees Taal, Olga Fink
Domain adaptation (DA) methods aim to address the domain shift problem by extracting domain invariant features.
no code implementations • 20 Dec 2022 • Wei Ma, Mengjie Zhao, Xiaofei Xie, Qiang Hu, Shangqing Liu, Jie Zhang, Wenhan Wang, Yang Liu
To further understand the code features learnt by these models, in this paper, we target two well-known representative code pre-trained models (i. e., CodeBERT and GraphCodeBERT) and devise a set of probing tasks for the syntax and semantics analysis.
no code implementations • 25 Oct 2022 • Junze Li, Mengjie Zhao, Yubo Xie, Antonis Maronikolakis, Pearl Pu, Hinrich Schütze
Humor is a magnetic component in everyday human interactions and communications.
no code implementations • Findings (ACL) 2022 • Sheng Liang, Mengjie Zhao, Hinrich Schütze
Recent research has made impressive progress in large-scale multimodal pre-training.
no code implementations • Findings (NAACL) 2022 • Mengjie Zhao, Fei Mi, Yasheng Wang, Minglei Li, Xin Jiang, Qun Liu, Hinrich Schütze
We propose LMTurk, a novel approach that treats few-shot learners as crowdsourcing workers.
1 code implementation • EMNLP 2021 • Mengjie Zhao, Hinrich Schütze
It has been shown for English that discrete and soft prompting perform strongly in few-shot learning with pretrained language models (PLMs).
no code implementations • ACL 2021 • Mengjie Zhao, Yi Zhu, Ehsan Shareghi, Ivan Vulić, Roi Reichart, Anna Korhonen, Hinrich Schütze
Few-shot crosslingual transfer has been shown to outperform its zero-shot counterpart with pretrained encoders like multilingual BERT.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Fei Mi, LiangWei Chen, Mengjie Zhao, Minlie Huang, Boi Faltings
Natural language generation (NLG) is an essential component of task-oriented dialog systems.
no code implementations • EMNLP 2020 • Mengjie Zhao, Tao Lin, Fei Mi, Martin Jaggi, Hinrich Schütze
We present an efficient method of utilizing pretrained language models, where we learn selective binary masks for pretrained weights in lieu of modifying them through finetuning.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Mengjie Zhao, Philipp Dufter, Yadollah Yaghoobzadeh, Hinrich Schütze
Pretrained language models have achieved a new state of the art on many NLP tasks, but there are still many open questions about how and why they work so well.
no code implementations • ACL 2019 • Mengjie Zhao, Hinrich Sch{\"u}tze
We present a new method for sentiment lexicon induction that is designed to be applicable to the entire range of typological diversity of the world{'}s languages.
no code implementations • 1 Nov 2018 • Philipp Dufter, Mengjie Zhao, Hinrich Schütze
A simple and effective context-based multilingual embedding learner is Levy et al. (2017)'s S-ID (sentence ID) method.
no code implementations • ACL 2018 • Philipp Dufter, Mengjie Zhao, Martin Schmitt, Alexander Fraser, Hinrich Schütze
We present a new method for estimating vector space representations of words: embedding learning by concept induction.