Search Results for author: Gábor Berend

Found 10 papers, 5 papers with code

Identifying the Importance of Content Overlap for Better Cross-lingual Embedding Mappings

no code implementations EMNLP (MRL) 2021 Réka Cserháti, Gábor Berend

In this work, we analyze the performance and properties of cross-lingual word embedding models created by mapping-based alignment methods.

Codenames as a Game of Co-occurrence Counting

1 code implementation CMCL (ACL) 2022 Réka Cserháti, Istvan Kollath, András Kicsi, Gábor Berend

This is a hard challenge even with today’s advanced language technology methods. In our study, we create spymaster agents using four types of relatedness measures that require only a raw text corpus to produce.

Quasi-Multitask Learning: an Efficient Surrogate for Obtaining Model Ensembles

no code implementations EMNLP (sustainlp) 2020 Norbert Kis-Szabó, Gábor Berend

We propose the technique of quasi-multitask learning (Q-MTL), a simple and easy to implement modification of standard multitask learning, in which the tasks to be modeled are identical.

Combating the Curse of Multilinguality in Cross-Lingual WSD by Aligning Sparse Contextualized Word Representations

1 code implementation NAACL 2022 Gábor Berend

In this paper, we advocate for using large pre-trained monolingual language models in cross lingual zero-shot word sense disambiguation (WSD) coupled with a contextualized mapping mechanism.

Dictionary Learning Word Sense Disambiguation

Masked Latent Semantic Modeling: an Efficient Pre-training Alternative to Masked Language Modeling

1 code implementation ACL Findings 2023 Gábor Berend

In this paper, we propose an alternative to the classic masked language modeling (MLM) pre-training paradigm, where the objective is altered from the reconstruction of the exact identity of randomly selected masked subwords to the prediction of their latent semantic properties.

Language Modelling Masked Language Modeling

Massively Multilingual Sparse Word Representations

1 code implementation ICLR 2020 Gábor Berend

Finally, we are releasing our multilingual sparse word representations for the 27 typologically diverse set of languages that we conducted our various experiments on.

Dependency Parsing Document Classification +1

Frustratingly easy quasi-multitask learning

no code implementations25 Sep 2019 Gábor Berend, Norbert Kis-Szabó

We propose the technique of quasi-multitask learning (Q-MTL), a simple and easy to implement modification of standard multitask learning, in which the tasks to be modeled are identical.

Sparse Coding of Neural Word Embeddings for Multilingual Sequence Labeling

no code implementations21 Dec 2016 Gábor Berend

In this paper we propose and carefully evaluate a sequence labeling framework which solely utilizes sparse indicator features derived from dense distributed word representations.

named-entity-recognition Named Entity Recognition +5

Cannot find the paper you are looking for? You can Submit a new open access paper.